var/home/core/zuul-output/0000755000175000017500000000000015067000160014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067007202015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004771406515067007173017721 0ustar rootrootSep 30 16:11:46 crc systemd[1]: Starting Kubernetes Kubelet... Sep 30 16:11:46 crc restorecon[4674]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 16:11:46 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 16:11:47 crc restorecon[4674]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 30 16:11:48 crc kubenswrapper[4796]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 16:11:48 crc kubenswrapper[4796]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 30 16:11:48 crc kubenswrapper[4796]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 16:11:48 crc kubenswrapper[4796]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 16:11:48 crc kubenswrapper[4796]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 30 16:11:48 crc kubenswrapper[4796]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.442472 4796 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447483 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447515 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447524 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447533 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447541 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447551 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447559 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447567 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447575 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447584 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447592 4796 feature_gate.go:330] unrecognized feature gate: Example Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447600 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447608 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447615 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447623 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447631 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447639 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447646 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447654 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447661 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447691 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447699 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447707 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447714 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447722 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447729 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447737 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447745 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447752 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447762 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447771 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447782 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447792 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447802 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447812 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447821 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447829 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447837 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447850 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447860 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447878 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447895 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447906 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447917 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447927 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447937 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447947 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447960 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.447971 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448016 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448028 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448038 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448049 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448059 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448070 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448080 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448117 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448129 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448143 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448161 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448175 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448187 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448199 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448209 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448220 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448229 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448245 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448258 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448269 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448280 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.448290 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449593 4796 flags.go:64] FLAG: --address="0.0.0.0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449625 4796 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449654 4796 flags.go:64] FLAG: --anonymous-auth="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449666 4796 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449677 4796 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449687 4796 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449699 4796 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449710 4796 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449719 4796 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449728 4796 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449738 4796 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449748 4796 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449756 4796 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449765 4796 flags.go:64] FLAG: --cgroup-root="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449774 4796 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449783 4796 flags.go:64] FLAG: --client-ca-file="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449791 4796 flags.go:64] FLAG: --cloud-config="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449800 4796 flags.go:64] FLAG: --cloud-provider="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449809 4796 flags.go:64] FLAG: --cluster-dns="[]" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449825 4796 flags.go:64] FLAG: --cluster-domain="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449834 4796 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449854 4796 flags.go:64] FLAG: --config-dir="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449864 4796 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449873 4796 flags.go:64] FLAG: --container-log-max-files="5" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449884 4796 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449893 4796 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449902 4796 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449912 4796 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449922 4796 flags.go:64] FLAG: --contention-profiling="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449931 4796 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449940 4796 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449949 4796 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449958 4796 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.449969 4796 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450004 4796 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450014 4796 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450022 4796 flags.go:64] FLAG: --enable-load-reader="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450032 4796 flags.go:64] FLAG: --enable-server="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450040 4796 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450059 4796 flags.go:64] FLAG: --event-burst="100" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450068 4796 flags.go:64] FLAG: --event-qps="50" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450077 4796 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450086 4796 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450095 4796 flags.go:64] FLAG: --eviction-hard="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450107 4796 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450115 4796 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450124 4796 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450133 4796 flags.go:64] FLAG: --eviction-soft="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450142 4796 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450151 4796 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450161 4796 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450170 4796 flags.go:64] FLAG: --experimental-mounter-path="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450180 4796 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450188 4796 flags.go:64] FLAG: --fail-swap-on="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450197 4796 flags.go:64] FLAG: --feature-gates="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450209 4796 flags.go:64] FLAG: --file-check-frequency="20s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450219 4796 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450241 4796 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450250 4796 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450260 4796 flags.go:64] FLAG: --healthz-port="10248" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450269 4796 flags.go:64] FLAG: --help="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450278 4796 flags.go:64] FLAG: --hostname-override="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450287 4796 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450296 4796 flags.go:64] FLAG: --http-check-frequency="20s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450305 4796 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450313 4796 flags.go:64] FLAG: --image-credential-provider-config="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450322 4796 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450331 4796 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450339 4796 flags.go:64] FLAG: --image-service-endpoint="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450348 4796 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450357 4796 flags.go:64] FLAG: --kube-api-burst="100" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450366 4796 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450375 4796 flags.go:64] FLAG: --kube-api-qps="50" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450383 4796 flags.go:64] FLAG: --kube-reserved="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450392 4796 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450401 4796 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450410 4796 flags.go:64] FLAG: --kubelet-cgroups="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450448 4796 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450457 4796 flags.go:64] FLAG: --lock-file="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450466 4796 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450475 4796 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450484 4796 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450497 4796 flags.go:64] FLAG: --log-json-split-stream="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450506 4796 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450515 4796 flags.go:64] FLAG: --log-text-split-stream="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450523 4796 flags.go:64] FLAG: --logging-format="text" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450532 4796 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450542 4796 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450552 4796 flags.go:64] FLAG: --manifest-url="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450561 4796 flags.go:64] FLAG: --manifest-url-header="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450572 4796 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450582 4796 flags.go:64] FLAG: --max-open-files="1000000" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450592 4796 flags.go:64] FLAG: --max-pods="110" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450613 4796 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450622 4796 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450632 4796 flags.go:64] FLAG: --memory-manager-policy="None" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450641 4796 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450650 4796 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450658 4796 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450667 4796 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450687 4796 flags.go:64] FLAG: --node-status-max-images="50" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450696 4796 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450706 4796 flags.go:64] FLAG: --oom-score-adj="-999" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450715 4796 flags.go:64] FLAG: --pod-cidr="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450724 4796 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450749 4796 flags.go:64] FLAG: --pod-manifest-path="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450758 4796 flags.go:64] FLAG: --pod-max-pids="-1" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450768 4796 flags.go:64] FLAG: --pods-per-core="0" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450777 4796 flags.go:64] FLAG: --port="10250" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450786 4796 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450795 4796 flags.go:64] FLAG: --provider-id="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450804 4796 flags.go:64] FLAG: --qos-reserved="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450813 4796 flags.go:64] FLAG: --read-only-port="10255" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450822 4796 flags.go:64] FLAG: --register-node="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450831 4796 flags.go:64] FLAG: --register-schedulable="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450839 4796 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450854 4796 flags.go:64] FLAG: --registry-burst="10" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450863 4796 flags.go:64] FLAG: --registry-qps="5" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450871 4796 flags.go:64] FLAG: --reserved-cpus="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450880 4796 flags.go:64] FLAG: --reserved-memory="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450891 4796 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450901 4796 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450910 4796 flags.go:64] FLAG: --rotate-certificates="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450920 4796 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450929 4796 flags.go:64] FLAG: --runonce="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450938 4796 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450947 4796 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450956 4796 flags.go:64] FLAG: --seccomp-default="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.450965 4796 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451010 4796 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451020 4796 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451030 4796 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451040 4796 flags.go:64] FLAG: --storage-driver-password="root" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451049 4796 flags.go:64] FLAG: --storage-driver-secure="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451058 4796 flags.go:64] FLAG: --storage-driver-table="stats" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451066 4796 flags.go:64] FLAG: --storage-driver-user="root" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451075 4796 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451085 4796 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451094 4796 flags.go:64] FLAG: --system-cgroups="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451102 4796 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451115 4796 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451124 4796 flags.go:64] FLAG: --tls-cert-file="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451133 4796 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451149 4796 flags.go:64] FLAG: --tls-min-version="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451158 4796 flags.go:64] FLAG: --tls-private-key-file="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451166 4796 flags.go:64] FLAG: --topology-manager-policy="none" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451175 4796 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451184 4796 flags.go:64] FLAG: --topology-manager-scope="container" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451193 4796 flags.go:64] FLAG: --v="2" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451205 4796 flags.go:64] FLAG: --version="false" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451215 4796 flags.go:64] FLAG: --vmodule="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451225 4796 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.451235 4796 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451495 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451507 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451519 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451528 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451538 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451547 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451555 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451563 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451571 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451578 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451586 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451595 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451619 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451627 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451635 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451642 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451650 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451658 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451665 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451673 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451680 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451688 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451696 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451703 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451711 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451719 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451727 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451735 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451743 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451750 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451758 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451766 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451776 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451785 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451795 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451806 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451815 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451824 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451832 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451841 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451849 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451857 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451865 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451873 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451880 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451888 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451895 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451903 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451922 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451930 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451938 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451946 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451954 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451961 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.451969 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452002 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452010 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452018 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452026 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452033 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452041 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452048 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452056 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452064 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452073 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452081 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452091 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452101 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452110 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452119 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.452127 4796 feature_gate.go:330] unrecognized feature gate: Example Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.452151 4796 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.468914 4796 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.468974 4796 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470697 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470734 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470743 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470753 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470762 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470771 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470779 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470789 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470801 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470813 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470824 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470832 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470840 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470848 4796 feature_gate.go:330] unrecognized feature gate: Example Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470859 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470869 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470877 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470886 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470895 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470904 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470912 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470920 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470929 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470936 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470944 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470952 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470960 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.470968 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471005 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471013 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471021 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471029 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471045 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471054 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471062 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471070 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471081 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471092 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471101 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471109 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471120 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471129 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471136 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471144 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471152 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471160 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471168 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471176 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471184 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471193 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471202 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471210 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471220 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471227 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471235 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471242 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471250 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471258 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471266 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471274 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471282 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471289 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471297 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471484 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471491 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471500 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471508 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471517 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471524 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471532 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471540 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.471553 4796 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471781 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471793 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471802 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471810 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471818 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471826 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471834 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471842 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471849 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471857 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471865 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471872 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471880 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471887 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471895 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471903 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471913 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471920 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471928 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471935 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471943 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471951 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471958 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471966 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.471998 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472009 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472019 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472028 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472037 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472045 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472053 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472061 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472069 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472077 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472084 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472095 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472104 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472116 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472124 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472132 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472140 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472150 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472159 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472167 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472175 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472183 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472190 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472198 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472206 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472214 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472221 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472229 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472237 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472245 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472252 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472260 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472267 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472277 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472285 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472293 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472301 4796 feature_gate.go:330] unrecognized feature gate: Example Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472308 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472316 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472324 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472335 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472345 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472353 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472361 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472370 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472379 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.472387 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.472398 4796 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.472634 4796 server.go:940] "Client rotation is on, will bootstrap in background" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.481294 4796 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.481425 4796 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.483177 4796 server.go:997] "Starting client certificate rotation" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.483231 4796 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.485275 4796 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-22 04:49:49.063262041 +0000 UTC Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.485403 4796 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1980h38m0.577863753s for next certificate rotation Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.515736 4796 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.518772 4796 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.540631 4796 log.go:25] "Validated CRI v1 runtime API" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.581842 4796 log.go:25] "Validated CRI v1 image API" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.584954 4796 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.592757 4796 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-30-16-07-10-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.592812 4796 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.624175 4796 manager.go:217] Machine: {Timestamp:2025-09-30 16:11:48.61843106 +0000 UTC m=+0.631709677 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:880be81d-8176-4d82-bbcc-7517e59a48bd BootID:6af019d7-e59f-45e8-81c8-64578033eb5e Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f3:41:0e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f3:41:0e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:33:c5:1f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f7:ee:56 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:fb:f8:c2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:90:90:1b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2e:67:f9:56:a7:2c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:e2:2b:28:a1:0a:96 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.624612 4796 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.625144 4796 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.628709 4796 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.630071 4796 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.630128 4796 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.630568 4796 topology_manager.go:138] "Creating topology manager with none policy" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.630589 4796 container_manager_linux.go:303] "Creating device plugin manager" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.631439 4796 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.631489 4796 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.633137 4796 state_mem.go:36] "Initialized new in-memory state store" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.633356 4796 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.637472 4796 kubelet.go:418] "Attempting to sync node with API server" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.637508 4796 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.637537 4796 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.637558 4796 kubelet.go:324] "Adding apiserver pod source" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.637577 4796 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.643121 4796 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.644619 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.644643 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.645572 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.645578 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.648167 4796 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.650362 4796 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652395 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652446 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652461 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652475 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652497 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652510 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652523 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652546 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652561 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652575 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652617 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.652632 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.653608 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.654424 4796 server.go:1280] "Started kubelet" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.655100 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.655737 4796 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.656314 4796 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.656586 4796 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 30 16:11:48 crc systemd[1]: Started Kubernetes Kubelet. Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658014 4796 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658088 4796 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658224 4796 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 17:08:45.038802111 +0000 UTC Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658280 4796 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 912h56m56.380526539s for next certificate rotation Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658345 4796 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658385 4796 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.658484 4796 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.658557 4796 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.659526 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.659666 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.660046 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="200ms" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.660953 4796 factory.go:55] Registering systemd factory Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.660975 4796 factory.go:221] Registration of the systemd container factory successfully Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.661351 4796 factory.go:153] Registering CRI-O factory Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.661391 4796 factory.go:221] Registration of the crio container factory successfully Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.661493 4796 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.661525 4796 factory.go:103] Registering Raw factory Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.661549 4796 manager.go:1196] Started watching for new ooms in manager Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.662358 4796 manager.go:319] Starting recovery of all containers Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.673316 4796 server.go:460] "Adding debug handlers to kubelet server" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.674912 4796 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.38:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a1b612ee86d49 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-30 16:11:48.654361929 +0000 UTC m=+0.667640496,LastTimestamp:2025-09-30 16:11:48.654361929 +0000 UTC m=+0.667640496,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682435 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682530 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682563 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682592 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682616 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682642 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682701 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682774 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682807 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682834 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682863 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682891 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682919 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.682950 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683019 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683054 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683083 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683109 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683135 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683161 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683185 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683215 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683239 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683266 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683293 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683317 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683349 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683377 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683405 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683430 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683459 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683485 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683625 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683659 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683685 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683711 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683746 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683776 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683803 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683832 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683856 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683891 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683926 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.683952 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684007 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684037 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684061 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684086 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684113 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684140 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684168 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684197 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684237 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684273 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684303 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684333 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684361 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684389 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684416 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684444 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684471 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684498 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684523 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684550 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684580 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684611 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684640 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684667 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684692 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684718 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684744 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684770 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684799 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684829 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684855 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684880 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684906 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684934 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.684959 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685025 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685059 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685090 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685116 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685140 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685166 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685192 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685219 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685245 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685271 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685297 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685323 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685353 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685382 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685408 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685501 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685538 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685569 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685598 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685627 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685655 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685684 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685710 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685737 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685767 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685804 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685834 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685863 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685890 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685919 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.685947 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686029 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686063 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686093 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686119 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686145 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686172 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686201 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686226 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686252 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686277 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686303 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686329 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686354 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686380 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686404 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686431 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686456 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686481 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686509 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686534 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686559 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686583 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686608 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686634 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686697 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686722 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686746 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686772 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686798 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686829 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686856 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686885 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686910 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686935 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.686963 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687050 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687082 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687107 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687132 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687161 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687185 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687214 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687240 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687267 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687295 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687318 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687344 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687369 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687394 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687419 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687447 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687475 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687499 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687526 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687548 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687572 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687594 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687624 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687647 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687674 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687699 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687723 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687747 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687769 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687793 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687818 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687844 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687870 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687894 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687918 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687941 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.687969 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688037 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688062 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688089 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688122 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688148 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688175 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688201 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688225 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688256 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688281 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688308 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688343 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688366 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688390 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.688414 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.691903 4796 manager.go:324] Recovery completed Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693490 4796 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693538 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693566 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693588 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693613 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693634 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693654 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693676 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693695 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693715 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693734 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693754 4796 reconstruct.go:97] "Volume reconstruction finished" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.693768 4796 reconciler.go:26] "Reconciler: start to sync state" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.701425 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.710847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.710917 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.710933 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.713149 4796 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.713173 4796 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.713199 4796 state_mem.go:36] "Initialized new in-memory state store" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.729826 4796 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.731852 4796 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.731913 4796 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.731954 4796 kubelet.go:2335] "Starting kubelet main sync loop" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.732033 4796 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 30 16:11:48 crc kubenswrapper[4796]: W0930 16:11:48.733146 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.733215 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.737944 4796 policy_none.go:49] "None policy: Start" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.738843 4796 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.738926 4796 state_mem.go:35] "Initializing new in-memory state store" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.758880 4796 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.805903 4796 manager.go:334] "Starting Device Plugin manager" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.806014 4796 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.806032 4796 server.go:79] "Starting device plugin registration server" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.806637 4796 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.806658 4796 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.807090 4796 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.807183 4796 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.807194 4796 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.822112 4796 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.832497 4796 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.832616 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.834159 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.834212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.834225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.834733 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.835436 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.835521 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.837103 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.837170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.837190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.837641 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.837714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.837731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.838175 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.838475 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.838515 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.842120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.842341 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.842374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.842388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.842676 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.842855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.843276 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.843409 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.843596 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.844483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.844522 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.844533 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.845416 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.845624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.845793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.846188 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.846310 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.846487 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.847336 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.847367 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.847380 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.847918 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.848178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.848330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.848677 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.848912 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.850296 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.850318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.850329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.861703 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="400ms" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897064 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897123 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897160 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897226 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897287 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897382 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897419 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897454 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897502 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897588 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897644 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897686 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.897739 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.906772 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.913499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.913541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.913554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.913583 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:11:48 crc kubenswrapper[4796]: E0930 16:11:48.914244 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.998857 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.998971 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999106 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999143 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999183 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999257 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999306 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999156 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999487 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999522 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999551 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999553 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999585 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999613 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999638 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999669 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999673 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999689 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999729 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999741 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:48 crc kubenswrapper[4796]: I0930 16:11:48.999763 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:48.999787 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:48.999795 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:48.999837 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:48.999893 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:48.999942 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.000030 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.000083 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.000136 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.114700 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.116297 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.116368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.116386 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.116447 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:11:49 crc kubenswrapper[4796]: E0930 16:11:49.117232 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.187773 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.201213 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.210724 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.237283 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: W0930 16:11:49.241249 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f99adf893310912630ba7e93a82c4fa9e35a4bd14b0d3a36f0f75777d397cc74 WatchSource:0}: Error finding container f99adf893310912630ba7e93a82c4fa9e35a4bd14b0d3a36f0f75777d397cc74: Status 404 returned error can't find the container with id f99adf893310912630ba7e93a82c4fa9e35a4bd14b0d3a36f0f75777d397cc74 Sep 30 16:11:49 crc kubenswrapper[4796]: W0930 16:11:49.242421 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f729faccb0602d855a27f8ef611ceb950fd77d4747bdae87c5046203acc36de6 WatchSource:0}: Error finding container f729faccb0602d855a27f8ef611ceb950fd77d4747bdae87c5046203acc36de6: Status 404 returned error can't find the container with id f729faccb0602d855a27f8ef611ceb950fd77d4747bdae87c5046203acc36de6 Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.246146 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:49 crc kubenswrapper[4796]: W0930 16:11:49.250161 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b19a6e7bf071be329b08f7b5a224275176b67548c7780298542922700eb7a6bb WatchSource:0}: Error finding container b19a6e7bf071be329b08f7b5a224275176b67548c7780298542922700eb7a6bb: Status 404 returned error can't find the container with id b19a6e7bf071be329b08f7b5a224275176b67548c7780298542922700eb7a6bb Sep 30 16:11:49 crc kubenswrapper[4796]: E0930 16:11:49.262740 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="800ms" Sep 30 16:11:49 crc kubenswrapper[4796]: W0930 16:11:49.269794 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e5dcd6e1d82aa9aa9ea84271257bb1b8f92b826eb2732ad7f4476bcadc6995e3 WatchSource:0}: Error finding container e5dcd6e1d82aa9aa9ea84271257bb1b8f92b826eb2732ad7f4476bcadc6995e3: Status 404 returned error can't find the container with id e5dcd6e1d82aa9aa9ea84271257bb1b8f92b826eb2732ad7f4476bcadc6995e3 Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.518250 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.519925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.519956 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.519967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.520005 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:11:49 crc kubenswrapper[4796]: E0930 16:11:49.520343 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.656010 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.737779 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5dcd6e1d82aa9aa9ea84271257bb1b8f92b826eb2732ad7f4476bcadc6995e3"} Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.739170 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"976061940acda9adcf57dc22993ce314b75534b337fc612c5e540e0912f37e1f"} Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.740813 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b19a6e7bf071be329b08f7b5a224275176b67548c7780298542922700eb7a6bb"} Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.741796 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f729faccb0602d855a27f8ef611ceb950fd77d4747bdae87c5046203acc36de6"} Sep 30 16:11:49 crc kubenswrapper[4796]: I0930 16:11:49.742599 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f99adf893310912630ba7e93a82c4fa9e35a4bd14b0d3a36f0f75777d397cc74"} Sep 30 16:11:49 crc kubenswrapper[4796]: W0930 16:11:49.814708 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:49 crc kubenswrapper[4796]: E0930 16:11:49.815027 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:49 crc kubenswrapper[4796]: W0930 16:11:49.827167 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:49 crc kubenswrapper[4796]: E0930 16:11:49.827236 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:50 crc kubenswrapper[4796]: E0930 16:11:50.064023 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="1.6s" Sep 30 16:11:50 crc kubenswrapper[4796]: W0930 16:11:50.155956 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:50 crc kubenswrapper[4796]: E0930 16:11:50.156060 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:50 crc kubenswrapper[4796]: W0930 16:11:50.185667 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:50 crc kubenswrapper[4796]: E0930 16:11:50.185793 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.321235 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.322851 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.322905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.322918 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.322946 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:11:50 crc kubenswrapper[4796]: E0930 16:11:50.323587 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.656589 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.749359 4796 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a" exitCode=0 Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.749465 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.749564 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.750895 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.750962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.751020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.752215 4796 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7" exitCode=0 Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.752263 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.752317 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.753904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.753947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.753965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.754447 4796 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac" exitCode=0 Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.754555 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.754653 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.755582 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.755631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.755643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.758792 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.758846 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.758857 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.758869 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.759086 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.760920 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.760963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.761018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.761180 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638" exitCode=0 Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.761234 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638"} Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.761311 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.762649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.762723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.762740 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.769919 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.776651 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.776692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:50 crc kubenswrapper[4796]: I0930 16:11:50.776722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.655880 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:51 crc kubenswrapper[4796]: E0930 16:11:51.665046 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="3.2s" Sep 30 16:11:51 crc kubenswrapper[4796]: W0930 16:11:51.687710 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Sep 30 16:11:51 crc kubenswrapper[4796]: E0930 16:11:51.687827 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.773679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.773733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.773744 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.773755 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.775032 4796 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd" exitCode=0 Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.775087 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.775213 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.776377 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.776427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.776444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.777664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"48af73cc5e378a96747a37927f934fb8f6591ff48a799d4cde3164a0a82053b5"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.777686 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.778940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.779030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.779046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.781994 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.782031 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.782046 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13"} Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.782052 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.782080 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.782952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.783009 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.783020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.783185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.783213 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.783224 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.924362 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.925809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.926554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.926590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:51 crc kubenswrapper[4796]: I0930 16:11:51.926628 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:11:51 crc kubenswrapper[4796]: E0930 16:11:51.928047 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.250216 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.788276 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029"} Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.788395 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.789606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.789660 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.789684 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.790882 4796 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723" exitCode=0 Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.790966 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.791080 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.791569 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.791608 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.791675 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.791089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723"} Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.795795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.795857 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.795876 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.796024 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.796091 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.796111 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.797674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.797715 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.797730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.798491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.798513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:52 crc kubenswrapper[4796]: I0930 16:11:52.798527 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.475199 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.802718 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.802782 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.802708 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc"} Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.802895 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a"} Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.802940 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c"} Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.802957 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7"} Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.804265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.804308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:53 crc kubenswrapper[4796]: I0930 16:11:53.804327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.604189 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.604479 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.606445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.606506 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.606527 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.735646 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.749921 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.813577 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e"} Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.813659 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.813699 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.813750 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.813770 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.813783 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815511 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815408 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815459 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815560 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:54 crc kubenswrapper[4796]: I0930 16:11:54.815590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.128213 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.130752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.130804 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.130823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.130857 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.250605 4796 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.250714 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.819838 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.819912 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.821106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.821135 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.821145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.821203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.821270 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.821293 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.825996 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.826227 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.827391 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.827442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:55 crc kubenswrapper[4796]: I0930 16:11:55.827464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:56 crc kubenswrapper[4796]: I0930 16:11:56.969199 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:11:56 crc kubenswrapper[4796]: I0930 16:11:56.969382 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:56 crc kubenswrapper[4796]: I0930 16:11:56.971049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:56 crc kubenswrapper[4796]: I0930 16:11:56.971117 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:56 crc kubenswrapper[4796]: I0930 16:11:56.971131 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:57 crc kubenswrapper[4796]: I0930 16:11:57.322633 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:11:57 crc kubenswrapper[4796]: I0930 16:11:57.322861 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:57 crc kubenswrapper[4796]: I0930 16:11:57.324258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:57 crc kubenswrapper[4796]: I0930 16:11:57.324292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:57 crc kubenswrapper[4796]: I0930 16:11:57.324302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:58 crc kubenswrapper[4796]: I0930 16:11:58.443476 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 30 16:11:58 crc kubenswrapper[4796]: I0930 16:11:58.443766 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:58 crc kubenswrapper[4796]: I0930 16:11:58.445414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:58 crc kubenswrapper[4796]: I0930 16:11:58.445469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:58 crc kubenswrapper[4796]: I0930 16:11:58.445483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:11:58 crc kubenswrapper[4796]: E0930 16:11:58.822666 4796 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 16:11:59 crc kubenswrapper[4796]: I0930 16:11:59.173705 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 30 16:11:59 crc kubenswrapper[4796]: I0930 16:11:59.174132 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:11:59 crc kubenswrapper[4796]: I0930 16:11:59.175793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:11:59 crc kubenswrapper[4796]: I0930 16:11:59.175844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:11:59 crc kubenswrapper[4796]: I0930 16:11:59.175855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:00 crc kubenswrapper[4796]: I0930 16:12:00.397425 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:12:00 crc kubenswrapper[4796]: I0930 16:12:00.397597 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:12:00 crc kubenswrapper[4796]: I0930 16:12:00.399266 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:00 crc kubenswrapper[4796]: I0930 16:12:00.399317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:00 crc kubenswrapper[4796]: I0930 16:12:00.399335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:02 crc kubenswrapper[4796]: I0930 16:12:02.441154 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 30 16:12:02 crc kubenswrapper[4796]: I0930 16:12:02.441237 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 16:12:02 crc kubenswrapper[4796]: I0930 16:12:02.449286 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 30 16:12:02 crc kubenswrapper[4796]: I0930 16:12:02.449348 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 16:12:03 crc kubenswrapper[4796]: I0930 16:12:03.494090 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]log ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]etcd ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/openshift.io-api-request-count-filter ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/openshift.io-startkubeinformers ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-apiserver-admission-initializer ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/generic-apiserver-start-informers ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/priority-and-fairness-config-consumer ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/priority-and-fairness-filter ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-apiextensions-informers ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-apiextensions-controllers ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/crd-informer-synced ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-system-namespaces-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-cluster-authentication-info-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-legacy-token-tracking-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-service-ip-repair-controllers ok Sep 30 16:12:03 crc kubenswrapper[4796]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/priority-and-fairness-config-producer ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/bootstrap-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/start-kube-aggregator-informers ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-status-local-available-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-status-remote-available-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-registration-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-wait-for-first-sync ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-discovery-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/kube-apiserver-autoregistration ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]autoregister-completion ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-openapi-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: [+]poststarthook/apiservice-openapiv3-controller ok Sep 30 16:12:03 crc kubenswrapper[4796]: livez check failed Sep 30 16:12:03 crc kubenswrapper[4796]: I0930 16:12:03.494154 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:12:05 crc kubenswrapper[4796]: I0930 16:12:05.251245 4796 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 16:12:05 crc kubenswrapper[4796]: I0930 16:12:05.251334 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.436283 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.438368 4796 trace.go:236] Trace[1822228969]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 16:11:52.562) (total time: 14875ms): Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[1822228969]: ---"Objects listed" error: 14875ms (16:12:07.438) Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[1822228969]: [14.875460002s] [14.875460002s] END Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.438800 4796 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.439559 4796 trace.go:236] Trace[1798841989]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 16:11:53.264) (total time: 14174ms): Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[1798841989]: ---"Objects listed" error: 14174ms (16:12:07.439) Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[1798841989]: [14.174756454s] [14.174756454s] END Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.439598 4796 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.440710 4796 trace.go:236] Trace[1829703138]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 16:11:55.996) (total time: 11444ms): Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[1829703138]: ---"Objects listed" error: 11444ms (16:12:07.440) Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[1829703138]: [11.444152857s] [11.444152857s] END Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.440752 4796 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.441193 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.441400 4796 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.442206 4796 trace.go:236] Trace[873734785]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 16:11:52.432) (total time: 15009ms): Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[873734785]: ---"Objects listed" error: 15009ms (16:12:07.442) Sep 30 16:12:07 crc kubenswrapper[4796]: Trace[873734785]: [15.00964938s] [15.00964938s] END Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.442238 4796 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.491362 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43898->192.168.126.11:17697: read: connection reset by peer" start-of-body= Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.491744 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43898->192.168.126.11:17697: read: connection reset by peer" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.491401 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.492129 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.649730 4796 apiserver.go:52] "Watching apiserver" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.654057 4796 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.654249 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.654550 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.654642 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.654730 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.654819 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.655036 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.655065 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.655157 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.655079 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.655212 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.656909 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.657153 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.657567 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.657716 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.657887 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.658088 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.658229 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.660110 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.660329 4796 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.662486 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.705267 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.716794 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.732840 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744132 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744176 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744192 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744207 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744225 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744242 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744258 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744282 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744307 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744324 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744380 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744396 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744414 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744432 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744454 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744474 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744490 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744508 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744525 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744542 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744557 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744571 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744586 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744587 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744599 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744689 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744716 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744744 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744769 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744794 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744817 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744874 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744896 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744917 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744942 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.744965 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745017 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745041 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745063 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745086 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745108 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745132 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745155 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745177 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745227 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745249 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745269 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745290 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745333 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745355 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745375 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745383 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745401 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745424 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745445 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745467 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745491 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745513 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745536 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745588 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745612 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745610 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745644 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745630 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745695 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745721 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745735 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745761 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745787 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745801 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745809 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745843 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745862 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745882 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745894 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745899 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745933 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745958 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.745957 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746008 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746023 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746035 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746051 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746055 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746084 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746102 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746121 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746136 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746159 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746176 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746192 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746207 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746226 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746254 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746273 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746293 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746317 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746332 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746349 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746364 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746379 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746396 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746412 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746428 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746459 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746481 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746500 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746521 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746542 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746563 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746580 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746595 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746612 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746628 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746646 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746662 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746717 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746737 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746756 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746773 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746790 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746806 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746822 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746841 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746857 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746876 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746892 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746910 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746927 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746943 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746960 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746993 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747010 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747028 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747048 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747083 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747104 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747134 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747149 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747167 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747183 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747201 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747218 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747235 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747252 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747270 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747286 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747303 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747320 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747336 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747353 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747369 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747388 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747404 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747422 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747446 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747473 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747491 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747509 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747576 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747594 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747611 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746225 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746307 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746597 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746603 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.746919 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.748322 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747142 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747374 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747404 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747491 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747635 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747761 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747814 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747874 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747861 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.747907 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.748176 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.748603 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.748628 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.748789 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.748807 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.749336 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:12:08.249299884 +0000 UTC m=+20.262578461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.749896 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750010 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750102 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750158 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750221 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750281 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750343 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750401 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750471 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750541 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750606 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750672 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750737 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750808 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750870 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750934 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751044 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751108 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751174 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751239 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751310 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751379 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751439 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751499 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751564 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751628 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751689 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751752 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751814 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751875 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751936 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752031 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752097 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752157 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752215 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752270 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752314 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752351 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753871 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753931 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754009 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754056 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754134 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754243 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754296 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754348 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754386 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754438 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754494 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754547 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754599 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754638 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754692 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754729 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754860 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754949 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755020 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755049 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755073 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755095 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755116 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755139 4796 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755161 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755183 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755203 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755224 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755247 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755268 4796 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755239 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755289 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755493 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755529 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755556 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755579 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755603 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755626 4796 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755647 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755666 4796 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755685 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755703 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755722 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755742 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755761 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755778 4796 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755799 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755821 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755840 4796 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750480 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750581 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750696 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750760 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750801 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750869 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750964 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.750948 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751161 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751347 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751358 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751425 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.751710 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.757745 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753036 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753143 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753313 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753625 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.753862 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754030 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.754963 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755483 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.755971 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.756304 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.756356 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.756405 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.756605 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.757072 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.757238 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.758026 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.758093 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.758138 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:08.258104037 +0000 UTC m=+20.271382564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.758370 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.752095 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759066 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759107 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759380 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759739 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759588 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759576 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.759844 4796 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.760814 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.761070 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.761363 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.761368 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.761395 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.761599 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.761594 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.762172 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.763897 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.762215 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.762454 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.762733 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.762868 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764090 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.763070 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.763170 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.763945 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764007 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764221 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764446 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764448 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764516 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.762245 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764543 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.764652 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764683 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.764703 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:08.264689016 +0000 UTC m=+20.277967553 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.765111 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.765187 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.765220 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.765373 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.765627 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.765808 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.766054 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.766058 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.766158 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.766259 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.766484 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.766624 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.767132 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.767276 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.767420 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.768289 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.768617 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.764937 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.768625 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.769023 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.769315 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.768824 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.770014 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.770207 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.771164 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.771500 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.771774 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.771800 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.772738 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.772768 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.772783 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.772839 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:08.272822369 +0000 UTC m=+20.286100916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.772957 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773166 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773327 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773361 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773524 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773651 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773705 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.773953 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.774416 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.774488 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.774906 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.774525 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.775038 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.775178 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.763774 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.775340 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.777059 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.777341 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.778452 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.778539 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.778762 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.779456 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.779573 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.779836 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.779952 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.780219 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.780370 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.780502 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.781241 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.781406 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.781668 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.781842 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.781935 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.782726 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.782856 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.783432 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.783608 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.784016 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.784541 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.784807 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.784819 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.785178 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.785096 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.786808 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.787090 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.786961 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.787262 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.787356 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.787383 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.787772 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.787954 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:07 crc kubenswrapper[4796]: E0930 16:12:07.788188 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:08.28816337 +0000 UTC m=+20.301441897 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.791727 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.792359 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.792362 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.792462 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.793585 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.793760 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.793765 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.794010 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.794259 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.794562 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.794590 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.795038 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.795071 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.795132 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.795174 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.795358 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.795641 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.796046 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.796151 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.796290 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.797098 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.797342 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.797829 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799062 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799167 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799369 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799365 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799410 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799780 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.799905 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.809191 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.810560 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.818540 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.822947 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.826530 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.855103 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856218 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856304 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856407 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856437 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856456 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856465 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856473 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856528 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856542 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856555 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856575 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856589 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856604 4796 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856616 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856628 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856641 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856653 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856668 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856680 4796 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856550 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856692 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856790 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856810 4796 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856827 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856843 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856856 4796 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856870 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856883 4796 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856896 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856909 4796 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856921 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856935 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856949 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.856961 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857021 4796 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857036 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857048 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857156 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857175 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857189 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857226 4796 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857466 4796 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857535 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857563 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029" exitCode=255 Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857575 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857600 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029"} Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857614 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857683 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857701 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857715 4796 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857732 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857745 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857758 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857771 4796 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857786 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857797 4796 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857810 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857822 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857833 4796 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857843 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857859 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857871 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857882 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857894 4796 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857906 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857919 4796 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857932 4796 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857946 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857957 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.857972 4796 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858007 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858022 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858064 4796 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858077 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858089 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858102 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858139 4796 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858152 4796 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858166 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858180 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858191 4796 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858205 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858218 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858229 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858242 4796 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858265 4796 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858276 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858288 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858300 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858311 4796 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858322 4796 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858333 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858343 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858355 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858366 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858376 4796 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858387 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858397 4796 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858408 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858427 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858438 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858448 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858460 4796 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858472 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858484 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858495 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858507 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858527 4796 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858539 4796 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858552 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858564 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858576 4796 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858587 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858598 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858611 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858627 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858640 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858653 4796 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858665 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858680 4796 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858693 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858707 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858718 4796 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858729 4796 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858740 4796 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858751 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858763 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858774 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858785 4796 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858797 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858809 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858820 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858832 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858844 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858855 4796 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858865 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858877 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858889 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858900 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858911 4796 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858923 4796 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858934 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858949 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858961 4796 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.858999 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859014 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859027 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859040 4796 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859052 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859062 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859072 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859084 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859095 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859106 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859119 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859132 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859144 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859157 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859170 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859181 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859194 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859206 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859219 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859241 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859261 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859280 4796 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859291 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859302 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.859313 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.870442 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.870929 4796 scope.go:117] "RemoveContainer" containerID="dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.870947 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.882022 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.894073 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.905224 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.915257 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.925023 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:07 crc kubenswrapper[4796]: I0930 16:12:07.978318 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 16:12:07 crc kubenswrapper[4796]: W0930 16:12:07.996622 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ca9ab5d2e03276529439dc1985c18c6482017314dcfaa0f29e5b96dda3329322 WatchSource:0}: Error finding container ca9ab5d2e03276529439dc1985c18c6482017314dcfaa0f29e5b96dda3329322: Status 404 returned error can't find the container with id ca9ab5d2e03276529439dc1985c18c6482017314dcfaa0f29e5b96dda3329322 Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.001394 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.011942 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 16:12:08 crc kubenswrapper[4796]: W0930 16:12:08.027614 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3be2973dcbf58de5e8cf215b6c9c0e114104b9ae8a0f5fe33ccccb3a483c28bd WatchSource:0}: Error finding container 3be2973dcbf58de5e8cf215b6c9c0e114104b9ae8a0f5fe33ccccb3a483c28bd: Status 404 returned error can't find the container with id 3be2973dcbf58de5e8cf215b6c9c0e114104b9ae8a0f5fe33ccccb3a483c28bd Sep 30 16:12:08 crc kubenswrapper[4796]: W0930 16:12:08.028748 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-becebdcf67b068653ee07866828a4f88997152c16b10ae499b94f7c85284e64e WatchSource:0}: Error finding container becebdcf67b068653ee07866828a4f88997152c16b10ae499b94f7c85284e64e: Status 404 returned error can't find the container with id becebdcf67b068653ee07866828a4f88997152c16b10ae499b94f7c85284e64e Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.264287 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.264474 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:12:09.264446513 +0000 UTC m=+21.277725040 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.265087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.265175 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.265218 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.265380 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:09.265363979 +0000 UTC m=+21.278642506 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.265404 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.265554 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:09.265545574 +0000 UTC m=+21.278824101 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.365706 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.365768 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.365896 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.365915 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.365927 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.366000 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:09.365960929 +0000 UTC m=+21.379239456 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.366060 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.366073 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.366081 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:08 crc kubenswrapper[4796]: E0930 16:12:08.366108 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:09.366100223 +0000 UTC m=+21.379378760 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.481476 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.494042 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.503766 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.513387 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.524115 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.536137 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.547625 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.560395 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.738159 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.739235 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.741643 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.743009 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.745021 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.746072 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.747305 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.749203 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.751022 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.752972 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.755830 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.755787 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.759110 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.760457 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.761586 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.763531 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.764580 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.766755 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.768091 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.769312 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.771895 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.773092 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.775662 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.776879 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.779806 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.780912 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.782673 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.783842 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.784464 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.784722 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.786244 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.786924 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.788155 4796 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.788371 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.790592 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.791831 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.792469 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.794435 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.795749 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.796502 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.797844 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.798941 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.799606 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.800860 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.802196 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.802431 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.803169 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.804275 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.805019 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.806315 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.808234 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.810066 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.811193 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.812221 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.813947 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.817079 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.818031 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.821189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.845694 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.860740 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.862148 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.862349 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca9ab5d2e03276529439dc1985c18c6482017314dcfaa0f29e5b96dda3329322"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.863723 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.865314 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.865897 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.867782 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.867841 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.867861 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"becebdcf67b068653ee07866828a4f88997152c16b10ae499b94f7c85284e64e"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.868521 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3be2973dcbf58de5e8cf215b6c9c0e114104b9ae8a0f5fe33ccccb3a483c28bd"} Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.869974 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.875193 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.886103 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.897447 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.908385 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.922515 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.938076 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.957017 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:08 crc kubenswrapper[4796]: I0930 16:12:08.975192 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.213122 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.230034 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.233228 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.237637 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.255785 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.271802 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.273016 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.273146 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.273201 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:12:11.273180522 +0000 UTC m=+23.286459069 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.273261 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.273278 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.273336 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:11.273313086 +0000 UTC m=+23.286591653 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.273414 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.273468 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:11.2734579 +0000 UTC m=+23.286736537 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.290136 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.306048 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.321197 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.342493 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.359803 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.373733 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.373797 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.373924 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.373945 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.373961 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.374051 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:11.374034548 +0000 UTC m=+23.387313075 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.374068 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.374120 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.374144 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.374241 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:11.374211913 +0000 UTC m=+23.387490480 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.376595 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.391402 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.414162 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.433632 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.448142 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.461524 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.486892 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.732375 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.732466 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.732495 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.732602 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:09 crc kubenswrapper[4796]: I0930 16:12:09.732645 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:09 crc kubenswrapper[4796]: E0930 16:12:09.732695 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.876498 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43"} Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.908952 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:10Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.925760 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:10Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.943423 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:10Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.958664 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:10Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.979191 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:10Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:10 crc kubenswrapper[4796]: I0930 16:12:10.994315 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:10Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.012065 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:11Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.034563 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:11Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.289948 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.290146 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.290193 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.290244 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:12:15.290196436 +0000 UTC m=+27.303475133 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.290314 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.290382 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:15.290362541 +0000 UTC m=+27.303641268 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.290393 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.290459 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:15.290444913 +0000 UTC m=+27.303723450 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.390887 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.390952 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391156 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391176 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391190 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391221 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391264 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:15.391243889 +0000 UTC m=+27.404522416 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391277 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391304 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.391399 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:15.391366333 +0000 UTC m=+27.404644900 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.733211 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.733281 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.733365 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.733493 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:11 crc kubenswrapper[4796]: I0930 16:12:11.733671 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:11 crc kubenswrapper[4796]: E0930 16:12:11.733779 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.256062 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.261052 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.273143 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.278398 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.296318 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.315221 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.351775 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.369524 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.386974 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.407908 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.427749 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.447252 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.469889 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.491634 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.509943 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.529041 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.559848 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.576843 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.593338 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: I0930 16:12:12.610713 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:12Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:12 crc kubenswrapper[4796]: E0930 16:12:12.900072 4796 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.732730 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.732861 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.732757 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.732967 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.733103 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.733335 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.841809 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.844704 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.844757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.844777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.844853 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.855230 4796 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.855651 4796 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.857313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.857374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.857389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.857411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.857425 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:13Z","lastTransitionTime":"2025-09-30T16:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.883356 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:13Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.889164 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.889233 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.889250 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.889283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.889347 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:13Z","lastTransitionTime":"2025-09-30T16:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.908488 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:13Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.913245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.913306 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.913323 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.913348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.913368 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:13Z","lastTransitionTime":"2025-09-30T16:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.930517 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:13Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.935416 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.935495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.935507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.935526 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.935539 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:13Z","lastTransitionTime":"2025-09-30T16:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.953536 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:13Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.958405 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.958457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.958470 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.958487 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.958503 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:13Z","lastTransitionTime":"2025-09-30T16:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.978386 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:13Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:13 crc kubenswrapper[4796]: E0930 16:12:13.978551 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.981229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.981298 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.981324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.981357 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:13 crc kubenswrapper[4796]: I0930 16:12:13.981380 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:13Z","lastTransitionTime":"2025-09-30T16:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.084007 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.084513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.084652 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.084810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.084949 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.188142 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.188223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.188246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.188277 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.188301 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.290804 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.290902 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.290925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.290957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.291011 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.393380 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.393431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.393446 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.393466 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.393477 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.495791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.495836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.495850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.495872 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.495883 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.597834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.597875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.597886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.597905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.597916 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.700990 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.701033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.701051 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.701069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.701082 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.802940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.802973 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.802999 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.803014 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.803022 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.858270 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-gpcn7"] Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.858606 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.859291 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-hldp6"] Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.859751 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862110 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862409 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862589 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862624 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862666 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862755 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862596 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.862799 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.875106 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.886533 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.907711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.907946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.907968 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.908010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.908025 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:14Z","lastTransitionTime":"2025-09-30T16:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.908819 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.920546 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923263 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-daemon-config\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923323 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-cni-bin\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923357 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-cnibin\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923377 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-multus-certs\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923398 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-system-cni-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923417 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-k8s-cni-cncf-io\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923437 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-cni-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923457 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-netns\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923475 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-kubelet\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923495 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z559s\" (UniqueName: \"kubernetes.io/projected/4f62bd4b-65ac-4b18-80ee-644b193c1912-kube-api-access-z559s\") pod \"node-resolver-gpcn7\" (UID: \"4f62bd4b-65ac-4b18-80ee-644b193c1912\") " pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923516 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-cni-multus\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923543 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-os-release\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923562 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-socket-dir-parent\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923580 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-conf-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-etc-kubernetes\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4f62bd4b-65ac-4b18-80ee-644b193c1912-hosts-file\") pod \"node-resolver-gpcn7\" (UID: \"4f62bd4b-65ac-4b18-80ee-644b193c1912\") " pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923649 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-cni-binary-copy\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923667 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-hostroot\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.923686 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6z6x\" (UniqueName: \"kubernetes.io/projected/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-kube-api-access-k6z6x\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.940683 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.953854 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.966624 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.982155 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:14 crc kubenswrapper[4796]: I0930 16:12:14.996063 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:14Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.010674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.010713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.010724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.010740 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.010750 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.011722 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.023365 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024661 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-cni-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024714 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-netns\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024747 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z559s\" (UniqueName: \"kubernetes.io/projected/4f62bd4b-65ac-4b18-80ee-644b193c1912-kube-api-access-z559s\") pod \"node-resolver-gpcn7\" (UID: \"4f62bd4b-65ac-4b18-80ee-644b193c1912\") " pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024772 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-cni-multus\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024796 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-kubelet\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024832 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-os-release\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024858 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-socket-dir-parent\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024880 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-conf-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024879 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-netns\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024953 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-etc-kubernetes\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024993 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-conf-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024993 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-socket-dir-parent\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025016 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-cni-multus\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.024902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-etc-kubernetes\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025002 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-kubelet\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025105 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-cni-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025058 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-os-release\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025185 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4f62bd4b-65ac-4b18-80ee-644b193c1912-hosts-file\") pod \"node-resolver-gpcn7\" (UID: \"4f62bd4b-65ac-4b18-80ee-644b193c1912\") " pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025216 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-hostroot\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025235 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6z6x\" (UniqueName: \"kubernetes.io/projected/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-kube-api-access-k6z6x\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025256 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-cni-binary-copy\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025283 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-hostroot\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025284 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-cni-bin\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025325 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-var-lib-cni-bin\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025357 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-daemon-config\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025420 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-cnibin\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025526 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-cnibin\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025472 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4f62bd4b-65ac-4b18-80ee-644b193c1912-hosts-file\") pod \"node-resolver-gpcn7\" (UID: \"4f62bd4b-65ac-4b18-80ee-644b193c1912\") " pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025554 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-multus-certs\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025583 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-multus-certs\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025583 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-system-cni-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025623 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-k8s-cni-cncf-io\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025753 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-host-run-k8s-cni-cncf-io\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.025799 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-system-cni-dir\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.026144 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-cni-binary-copy\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.026195 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-multus-daemon-config\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.033678 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.042445 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z559s\" (UniqueName: \"kubernetes.io/projected/4f62bd4b-65ac-4b18-80ee-644b193c1912-kube-api-access-z559s\") pod \"node-resolver-gpcn7\" (UID: \"4f62bd4b-65ac-4b18-80ee-644b193c1912\") " pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.049628 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6z6x\" (UniqueName: \"kubernetes.io/projected/b4cf8ea3-7aa4-4dfc-8443-347fd3898f15-kube-api-access-k6z6x\") pod \"multus-hldp6\" (UID: \"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\") " pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.060250 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.072927 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.085399 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.098054 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.112554 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.113359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.113404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.113414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.113432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.113442 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.124596 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.138120 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.151035 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.165635 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.171922 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-gpcn7" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.176159 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hldp6" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.217670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.218003 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.218014 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.218031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.218043 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.247092 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jskcj"] Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.247813 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.249008 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-8bbws"] Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.249248 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c6n4h"] Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.249576 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.253405 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.253715 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.254914 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.256878 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.258260 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.258840 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.259011 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.259051 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.259371 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.259575 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.259586 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.260045 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.260354 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.260447 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.260472 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.265459 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.293149 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.319968 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.322016 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.322136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.322232 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.322310 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.322382 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329131 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329470 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-var-lib-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmcms\" (UniqueName: \"kubernetes.io/projected/f7df6e69-ff3a-484a-b440-8afe12028b40-kube-api-access-jmcms\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329684 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-ovn-kubernetes\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329771 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-config\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329860 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cgzg\" (UniqueName: \"kubernetes.io/projected/414dbf05-2d31-4c78-aad6-6c8f890f87ae-kube-api-access-4cgzg\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.329928 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-kubelet\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330025 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-systemd\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330101 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-netd\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.330164 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:12:23.330134896 +0000 UTC m=+35.343413433 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-node-log\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330264 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330289 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-etc-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330312 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-os-release\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330331 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-log-socket\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330351 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-env-overrides\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330390 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/670c655e-3953-4773-84ef-19c678d482f9-mcd-auth-proxy-config\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330413 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-netns\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330436 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/670c655e-3953-4773-84ef-19c678d482f9-rootfs\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330455 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/670c655e-3953-4773-84ef-19c678d482f9-proxy-tls\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330479 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330501 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330534 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330572 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-systemd-units\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cnibin\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330618 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330640 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330662 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-bin\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330683 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vch6n\" (UniqueName: \"kubernetes.io/projected/670c655e-3953-4773-84ef-19c678d482f9-kube-api-access-vch6n\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330704 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-system-cni-dir\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330722 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-slash\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330743 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-script-lib\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330763 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330790 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7df6e69-ff3a-484a-b440-8afe12028b40-ovn-node-metrics-cert\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.330813 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-ovn\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.331173 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.331298 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:23.331280329 +0000 UTC m=+35.344558846 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.331796 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.331900 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:23.331891396 +0000 UTC m=+35.345169923 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.355284 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.378691 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.419654 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.424296 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.424327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.424335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.424348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.424356 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432374 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432431 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7df6e69-ff3a-484a-b440-8afe12028b40-ovn-node-metrics-cert\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432453 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-ovn\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432476 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-var-lib-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432495 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmcms\" (UniqueName: \"kubernetes.io/projected/f7df6e69-ff3a-484a-b440-8afe12028b40-kube-api-access-jmcms\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-ovn-kubernetes\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432529 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432538 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-config\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432593 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-ovn\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432617 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cgzg\" (UniqueName: \"kubernetes.io/projected/414dbf05-2d31-4c78-aad6-6c8f890f87ae-kube-api-access-4cgzg\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432696 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-kubelet\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432718 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-systemd\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432735 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-netd\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432761 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432779 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-etc-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432794 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-node-log\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-os-release\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432828 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-log-socket\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432862 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-env-overrides\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432891 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/670c655e-3953-4773-84ef-19c678d482f9-mcd-auth-proxy-config\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432907 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-netns\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-ovn-kubernetes\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432932 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432953 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/670c655e-3953-4773-84ef-19c678d482f9-rootfs\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432959 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-var-lib-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.432989 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/670c655e-3953-4773-84ef-19c678d482f9-proxy-tls\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433013 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433033 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433041 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-os-release\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433069 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-kubelet\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433079 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433097 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-systemd\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433103 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-systemd-units\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-netd\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433152 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cnibin\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433159 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-systemd-units\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433172 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433197 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/670c655e-3953-4773-84ef-19c678d482f9-rootfs\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433201 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vch6n\" (UniqueName: \"kubernetes.io/projected/670c655e-3953-4773-84ef-19c678d482f9-kube-api-access-vch6n\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433222 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-system-cni-dir\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433241 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-slash\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433258 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-bin\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433278 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-script-lib\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.433300 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.433339 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.433358 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.433428 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:23.433405492 +0000 UTC m=+35.446684089 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433804 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-config\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433806 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-log-socket\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.433897 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-script-lib\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434080 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-system-cni-dir\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434113 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-slash\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434135 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-bin\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434270 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-etc-openvswitch\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434281 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-node-log\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434364 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434393 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cnibin\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434421 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.434504 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434674 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/670c655e-3953-4773-84ef-19c678d482f9-mcd-auth-proxy-config\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.434748 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.434809 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.434903 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:23.434888865 +0000 UTC m=+35.448167392 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434514 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-env-overrides\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434807 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.434965 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/414dbf05-2d31-4c78-aad6-6c8f890f87ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.435215 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-netns\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.436278 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/670c655e-3953-4773-84ef-19c678d482f9-proxy-tls\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.439523 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7df6e69-ff3a-484a-b440-8afe12028b40-ovn-node-metrics-cert\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.446633 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.460176 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vch6n\" (UniqueName: \"kubernetes.io/projected/670c655e-3953-4773-84ef-19c678d482f9-kube-api-access-vch6n\") pod \"machine-config-daemon-8bbws\" (UID: \"670c655e-3953-4773-84ef-19c678d482f9\") " pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.468975 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cgzg\" (UniqueName: \"kubernetes.io/projected/414dbf05-2d31-4c78-aad6-6c8f890f87ae-kube-api-access-4cgzg\") pod \"multus-additional-cni-plugins-jskcj\" (UID: \"414dbf05-2d31-4c78-aad6-6c8f890f87ae\") " pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.473840 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmcms\" (UniqueName: \"kubernetes.io/projected/f7df6e69-ff3a-484a-b440-8afe12028b40-kube-api-access-jmcms\") pod \"ovnkube-node-c6n4h\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.476446 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.490450 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.511284 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.526964 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.527017 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.527026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.527040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.527050 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.528454 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.541260 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.552929 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.565038 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.575118 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.589231 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.593415 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jskcj" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.607409 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:12:15 crc kubenswrapper[4796]: W0930 16:12:15.607777 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod414dbf05_2d31_4c78_aad6_6c8f890f87ae.slice/crio-fbbe70b83cba099ebd53b41fc0b1e8e8b9d46b15412640209e37abd1c594d92f WatchSource:0}: Error finding container fbbe70b83cba099ebd53b41fc0b1e8e8b9d46b15412640209e37abd1c594d92f: Status 404 returned error can't find the container with id fbbe70b83cba099ebd53b41fc0b1e8e8b9d46b15412640209e37abd1c594d92f Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.619079 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.619710 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.630964 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.631087 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.631099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.631114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.631125 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.645412 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: W0930 16:12:15.648328 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7df6e69_ff3a_484a_b440_8afe12028b40.slice/crio-7f6acb784913fb54b9caa46e850bbf8d7467eabc4a7306815d95e9a2951b8925 WatchSource:0}: Error finding container 7f6acb784913fb54b9caa46e850bbf8d7467eabc4a7306815d95e9a2951b8925: Status 404 returned error can't find the container with id 7f6acb784913fb54b9caa46e850bbf8d7467eabc4a7306815d95e9a2951b8925 Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.658373 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.689400 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.711847 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.728680 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.732198 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.732199 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.732310 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.732377 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.732198 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:15 crc kubenswrapper[4796]: E0930 16:12:15.732516 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.733899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.733925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.733936 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.733950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.733963 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.741400 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.754147 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.769242 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.782281 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.835868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.835900 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.835911 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.835926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.835936 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.901025 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-gpcn7" event={"ID":"4f62bd4b-65ac-4b18-80ee-644b193c1912","Type":"ContainerStarted","Data":"6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.901086 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-gpcn7" event={"ID":"4f62bd4b-65ac-4b18-80ee-644b193c1912","Type":"ContainerStarted","Data":"a736c0f2a21e193b955eb239cbd02166a587914620ddcab7303f74eb3b198a03"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.903075 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" exitCode=0 Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.903105 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.903219 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"7f6acb784913fb54b9caa46e850bbf8d7467eabc4a7306815d95e9a2951b8925"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.906646 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.906695 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"8c81b2e3b259b2a75736f1c977c4d3a0dc396441a717787656d3a312ff280087"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.908718 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerStarted","Data":"d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.912315 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerStarted","Data":"fbbe70b83cba099ebd53b41fc0b1e8e8b9d46b15412640209e37abd1c594d92f"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.912382 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hldp6" event={"ID":"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15","Type":"ContainerStarted","Data":"1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.912405 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hldp6" event={"ID":"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15","Type":"ContainerStarted","Data":"056142ce2beb9d41a875dbf3a7bb75798eb454f9cb1eb48880c92dbdb0088e6d"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.918462 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.936372 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.942512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.942550 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.942560 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.942579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.942591 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:15Z","lastTransitionTime":"2025-09-30T16:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.957145 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:15 crc kubenswrapper[4796]: I0930 16:12:15.990136 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.003071 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.018113 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.033176 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.045674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.045846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.045930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.046083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.046186 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.046947 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.070147 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.085412 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.100282 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.119131 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.139875 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.148542 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.148612 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.148630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.148653 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.148668 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.154579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.168644 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.180582 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.191545 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.208311 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.220656 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.235850 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.249151 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.250690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.250927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.251506 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.251532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.251548 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.268189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.289376 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.307963 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.325510 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.339594 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.355123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.355174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.355185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.355204 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.355215 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.371817 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.409947 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.457604 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.457648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.457659 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.457677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.457687 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.560007 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.560060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.560076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.560100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.560117 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.662608 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.662682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.662696 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.662713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.662727 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.765350 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.765395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.765404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.765422 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.765432 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.867529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.867559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.867568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.867582 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.867591 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.919563 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.919602 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.919612 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.919622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.919631 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.919641 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.920908 4796 generic.go:334] "Generic (PLEG): container finished" podID="414dbf05-2d31-4c78-aad6-6c8f890f87ae" containerID="d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb" exitCode=0 Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.920962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerDied","Data":"d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.922882 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.938589 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.953091 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.971032 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.973795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.974071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.974149 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.974225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.974287 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:16Z","lastTransitionTime":"2025-09-30T16:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.982338 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:16 crc kubenswrapper[4796]: I0930 16:12:16.995566 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:16Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.006786 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.016753 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.027463 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.047000 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.059462 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.072737 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.077658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.077704 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.077753 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.077775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.077794 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.083726 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.094511 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.115188 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.126378 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.138820 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.150389 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.171052 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.180446 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.180502 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.180519 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.180544 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.180560 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.184238 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.212904 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.250069 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.282574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.282612 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.282624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.282642 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.282655 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.297895 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.337368 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.375639 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.385504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.385544 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.385552 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.385565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.385573 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.412176 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.453502 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.488178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.488223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.488233 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.488252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.488262 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.493823 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.501802 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xtn5f"] Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.502402 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.520561 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.541596 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.557643 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7530dcc2-0e7d-41df-b43c-13a1a28cee98-serviceca\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.557739 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530dcc2-0e7d-41df-b43c-13a1a28cee98-host\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.557766 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s8nb\" (UniqueName: \"kubernetes.io/projected/7530dcc2-0e7d-41df-b43c-13a1a28cee98-kube-api-access-2s8nb\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.560857 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.581315 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.591312 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.591363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.591375 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.591395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.591409 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.610695 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.649755 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.659329 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530dcc2-0e7d-41df-b43c-13a1a28cee98-host\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.659390 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s8nb\" (UniqueName: \"kubernetes.io/projected/7530dcc2-0e7d-41df-b43c-13a1a28cee98-kube-api-access-2s8nb\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.659436 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7530dcc2-0e7d-41df-b43c-13a1a28cee98-serviceca\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.659502 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530dcc2-0e7d-41df-b43c-13a1a28cee98-host\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.660913 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7530dcc2-0e7d-41df-b43c-13a1a28cee98-serviceca\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.693853 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.694426 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.694443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.694473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.694491 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.707123 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s8nb\" (UniqueName: \"kubernetes.io/projected/7530dcc2-0e7d-41df-b43c-13a1a28cee98-kube-api-access-2s8nb\") pod \"node-ca-xtn5f\" (UID: \"7530dcc2-0e7d-41df-b43c-13a1a28cee98\") " pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.715144 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.733124 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.733146 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:17 crc kubenswrapper[4796]: E0930 16:12:17.733263 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:17 crc kubenswrapper[4796]: E0930 16:12:17.733363 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.733226 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:17 crc kubenswrapper[4796]: E0930 16:12:17.733468 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.749286 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.789120 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.796751 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.796792 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.796805 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.796820 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.796829 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.832225 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.837524 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xtn5f" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.875482 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.899553 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.899596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.899607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.899628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.899643 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:17Z","lastTransitionTime":"2025-09-30T16:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.912375 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.927344 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xtn5f" event={"ID":"7530dcc2-0e7d-41df-b43c-13a1a28cee98","Type":"ContainerStarted","Data":"f49c536f6e90ce11f5c4a7f62c8b76258a548faaca7c5ea7e5367b9c1e2f9aef"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.929710 4796 generic.go:334] "Generic (PLEG): container finished" podID="414dbf05-2d31-4c78-aad6-6c8f890f87ae" containerID="c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52" exitCode=0 Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.929913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerDied","Data":"c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52"} Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.948942 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:17 crc kubenswrapper[4796]: I0930 16:12:17.993071 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:17Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.002848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.002897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.002910 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.002930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.002946 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.036804 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.071225 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.108334 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.108418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.108434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.108461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.108478 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.108462 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.149498 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.188851 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.211135 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.211168 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.211179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.211193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.211204 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.231486 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.269784 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.309616 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.313387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.313424 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.313438 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.313456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.313467 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.348251 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.390914 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.419684 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.419751 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.419769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.419795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.419819 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.429942 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.470384 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.509041 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.522823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.522865 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.522876 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.522892 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.522903 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.555829 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.601131 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.625155 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.625621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.625633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.625649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.625661 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.630900 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.670107 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.711125 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.728378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.728414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.728425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.728442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.728453 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.749242 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.798941 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.831368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.831451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.831481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.831516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.831541 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.836101 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.900217 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.920580 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.933709 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.933748 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.933778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.933797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.933808 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:18Z","lastTransitionTime":"2025-09-30T16:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.938520 4796 generic.go:334] "Generic (PLEG): container finished" podID="414dbf05-2d31-4c78-aad6-6c8f890f87ae" containerID="79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b" exitCode=0 Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.938587 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerDied","Data":"79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.948340 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.950335 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xtn5f" event={"ID":"7530dcc2-0e7d-41df-b43c-13a1a28cee98","Type":"ContainerStarted","Data":"a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473"} Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.954807 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:18 crc kubenswrapper[4796]: I0930 16:12:18.990456 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:18Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.031224 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.036175 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.036228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.036238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.036257 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.036269 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.072017 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.109396 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.139871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.139934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.139957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.140025 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.140046 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.150806 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.189037 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.236433 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.243351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.243400 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.243414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.243436 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.243453 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.270515 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.316531 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.347203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.347245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.347256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.347275 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.347291 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.361393 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.392677 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.442317 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.449779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.449806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.449816 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.449832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.449844 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.487128 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.510471 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.552437 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.554722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.554762 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.554779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.554803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.554821 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.590907 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.633049 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.657945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.658030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.658054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.658078 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.658102 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.675774 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.710307 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.732753 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.732945 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:19 crc kubenswrapper[4796]: E0930 16:12:19.733524 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.734144 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:19 crc kubenswrapper[4796]: E0930 16:12:19.734304 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:19 crc kubenswrapper[4796]: E0930 16:12:19.734452 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.755135 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.760587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.760793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.760926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.761099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.761816 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.793426 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.829741 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.865214 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.865275 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.865293 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.865319 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.865337 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.881148 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.916018 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.957385 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.958699 4796 generic.go:334] "Generic (PLEG): container finished" podID="414dbf05-2d31-4c78-aad6-6c8f890f87ae" containerID="a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6" exitCode=0 Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.958762 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerDied","Data":"a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.971174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.971240 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.971262 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.971292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.971345 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:19Z","lastTransitionTime":"2025-09-30T16:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:19 crc kubenswrapper[4796]: I0930 16:12:19.993198 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:19Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.047348 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.072163 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.074333 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.074365 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.074376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.074696 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.074718 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.111878 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.149055 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.176903 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.176944 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.176954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.176970 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.176995 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.189788 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.232462 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.272373 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.279152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.279192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.279207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.279228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.279244 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.312942 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.351775 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.381770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.381806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.381815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.381830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.381841 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.402270 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.438459 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.485009 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.485064 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.485082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.485107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.485124 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.506557 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.527809 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.555540 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.588685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.588724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.588736 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.588751 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.588761 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.591519 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.637175 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:20Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.691620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.691684 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.691701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.691724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.691740 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.794557 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.794632 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.794648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.794668 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.794682 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.897757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.897827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.897848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.897875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.897893 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:20Z","lastTransitionTime":"2025-09-30T16:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:20 crc kubenswrapper[4796]: I0930 16:12:20.967949 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerStarted","Data":"de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.001454 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.001524 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.001542 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.001568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.001586 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.003751 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.030459 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.045949 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.064470 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.085199 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.104628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.104674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.104686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.104707 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.104722 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.107775 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.122728 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.141779 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.153186 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.173562 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.189596 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.208851 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.208911 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.208931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.208959 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.209008 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.210712 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.228772 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.246665 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.275889 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:21Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.312720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.312783 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.312809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.312853 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.312878 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.416029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.416092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.416115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.416148 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.416169 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.518868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.518931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.518942 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.518963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.518990 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.622163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.622223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.622237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.622261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.622276 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.725664 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.725720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.725734 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.725757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.725770 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.733064 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.733120 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.733073 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:21 crc kubenswrapper[4796]: E0930 16:12:21.733254 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:21 crc kubenswrapper[4796]: E0930 16:12:21.733372 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:21 crc kubenswrapper[4796]: E0930 16:12:21.733560 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.828624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.828681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.828693 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.828717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.828732 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.932003 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.932051 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.932063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.932084 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.932098 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:21Z","lastTransitionTime":"2025-09-30T16:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.979018 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5"} Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.984210 4796 generic.go:334] "Generic (PLEG): container finished" podID="414dbf05-2d31-4c78-aad6-6c8f890f87ae" containerID="de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9" exitCode=0 Sep 30 16:12:21 crc kubenswrapper[4796]: I0930 16:12:21.984281 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerDied","Data":"de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.018796 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.034913 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.034967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.034991 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.035013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.035029 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.039772 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.061758 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.080585 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.100562 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.119220 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.138702 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.138780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.138796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.138815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.138827 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.139805 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.157427 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.178099 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.196112 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.211981 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.230523 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.241271 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.241318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.241330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.241353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.241369 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.244581 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.260067 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.288626 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.303765 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.317318 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.330117 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.345125 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.345192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.345207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.345234 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.345250 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.356014 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.382628 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.399292 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.417144 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.435415 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.448744 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.448795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.448809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.448832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.448845 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.455107 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.469930 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.483675 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.497863 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.510896 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.527455 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.540984 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:22Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.552621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.552687 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.552707 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.553200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.553261 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.656966 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.657026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.657038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.657054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.657064 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.759718 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.759768 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.759785 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.759811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.759828 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.864164 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.864233 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.864252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.864279 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.864299 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.967868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.967932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.967949 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.967973 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.968034 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:22Z","lastTransitionTime":"2025-09-30T16:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.993510 4796 generic.go:334] "Generic (PLEG): container finished" podID="414dbf05-2d31-4c78-aad6-6c8f890f87ae" containerID="fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525" exitCode=0 Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.993591 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerDied","Data":"fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525"} Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.993762 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.993951 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:22 crc kubenswrapper[4796]: I0930 16:12:22.994356 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.017491 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.041353 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.041556 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.047809 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.063374 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.076025 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.076083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.076101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.076129 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.076183 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.079609 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.097614 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.110451 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.131613 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.147757 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.168153 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.178734 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.178802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.178864 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.178894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.178916 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.183790 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.208599 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.234045 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.252282 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.267015 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.281757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.281807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.281823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.281848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.281866 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.285311 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.305443 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.322102 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.336665 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.356505 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.369516 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.385466 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.385533 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.385546 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.385569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.385585 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.386881 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.422665 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.430527 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.430676 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:12:39.430654357 +0000 UTC m=+51.443932894 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.430739 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.430825 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.430905 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.430956 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:39.430946426 +0000 UTC m=+51.444224963 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.430954 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.431057 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:39.431036938 +0000 UTC m=+51.444315505 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.443882 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.470670 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.482617 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.487412 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.487448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.487457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.487473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.487483 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.496670 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.510599 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.531830 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532170 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532195 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532207 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532260 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:39.532244446 +0000 UTC m=+51.545522973 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.532311 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532492 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532512 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532530 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.532594 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:39.532578855 +0000 UTC m=+51.545857392 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.549655 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.590244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.590302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.590320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.590346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.590366 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.602082 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.633656 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:23Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.693317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.693420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.693441 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.693588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.693624 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.733080 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.733193 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.733227 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.733322 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.733108 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:23 crc kubenswrapper[4796]: E0930 16:12:23.733782 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.797612 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.797680 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.797692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.797717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.797730 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.901080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.901164 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.901203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.901243 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:23 crc kubenswrapper[4796]: I0930 16:12:23.901268 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:23Z","lastTransitionTime":"2025-09-30T16:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.002771 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" event={"ID":"414dbf05-2d31-4c78-aad6-6c8f890f87ae","Type":"ContainerStarted","Data":"173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.002936 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.003431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.003485 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.003507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.003536 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.003559 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.023448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.023516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.023531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.023552 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.023569 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.032473 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: E0930 16:12:24.048106 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.052537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.052611 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.052636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.052669 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.052693 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.053702 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.075185 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: E0930 16:12:24.076224 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.080507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.080547 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.080558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.080574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.080585 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.090573 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: E0930 16:12:24.104390 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.108393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.108448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.108457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.108477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.108490 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.113532 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: E0930 16:12:24.132161 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.134429 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.136715 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.136760 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.136774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.136795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.136809 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.148215 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: E0930 16:12:24.148586 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: E0930 16:12:24.148693 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.150236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.150295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.150314 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.150333 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.150345 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.161672 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.183559 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.208520 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.225925 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.239759 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.252782 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.253482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.253622 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.253722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.253817 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.253885 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.267259 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.281557 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:24Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.356238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.356568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.356786 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.356877 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.356960 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.461037 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.461095 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.461110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.461134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.461145 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.564537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.564601 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.564621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.564645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.564665 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.668857 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.668946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.668963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.669001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.669014 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.771497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.771573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.771655 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.771690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.771709 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.874866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.874929 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.874951 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.874982 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.875025 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.978544 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.978629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.978649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.978678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:24 crc kubenswrapper[4796]: I0930 16:12:24.978698 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:24Z","lastTransitionTime":"2025-09-30T16:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.009605 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/0.log" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.014593 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5" exitCode=1 Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.014689 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.015524 4796 scope.go:117] "RemoveContainer" containerID="302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.034465 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.069879 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.081929 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.081997 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.082040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.082066 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.082084 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.093865 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.111901 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.131203 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.151766 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.173623 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.184717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.184769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.184782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.184804 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.184819 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.187452 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.199581 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.215306 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.229620 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.253535 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:24.709145 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:24.709163 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 16:12:24.709179 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 16:12:24.709194 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 16:12:24.709237 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 16:12:24.709255 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:24.709260 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 16:12:24.709305 6077 factory.go:656] Stopping watch factory\\\\nI0930 16:12:24.709331 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 16:12:24.709334 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 16:12:24.709351 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 16:12:24.709344 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 16:12:24.709367 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:24.709376 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 16:12:24.709396 6077 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.269018 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.284143 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.287092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.287134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.287143 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.287157 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.287170 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.304677 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.390031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.390097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.390116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.390147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.390169 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.493274 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.493334 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.493345 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.493362 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.493374 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.596835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.597382 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.597532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.597674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.597799 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.700120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.700194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.700217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.700248 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.700269 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.733282 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.733344 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.733430 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:25 crc kubenswrapper[4796]: E0930 16:12:25.733522 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:25 crc kubenswrapper[4796]: E0930 16:12:25.733777 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:25 crc kubenswrapper[4796]: E0930 16:12:25.733892 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.804092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.804158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.804185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.804218 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.804248 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.832209 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.851888 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.870152 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.886797 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.907045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.907091 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.907105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.907122 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.907134 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:25Z","lastTransitionTime":"2025-09-30T16:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.914957 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:24.709145 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:24.709163 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 16:12:24.709179 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 16:12:24.709194 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 16:12:24.709237 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 16:12:24.709255 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:24.709260 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 16:12:24.709305 6077 factory.go:656] Stopping watch factory\\\\nI0930 16:12:24.709331 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 16:12:24.709334 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 16:12:24.709351 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 16:12:24.709344 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 16:12:24.709367 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:24.709376 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 16:12:24.709396 6077 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.943560 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.973174 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:25 crc kubenswrapper[4796]: I0930 16:12:25.994817 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:25Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.010097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.010160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.010175 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.010201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.010224 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.012453 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.021874 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/0.log" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.025520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.025637 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.033122 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.052165 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.078114 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.094751 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.107629 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.112475 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.112525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.112535 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.112551 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.112560 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.132742 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.150020 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.166136 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.180826 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.193619 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.206024 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.215589 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.215636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.215645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.215661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.215672 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.220524 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.235608 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.248080 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.263435 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.279315 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.294187 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.316438 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:24.709145 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:24.709163 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 16:12:24.709179 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 16:12:24.709194 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 16:12:24.709237 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 16:12:24.709255 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:24.709260 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 16:12:24.709305 6077 factory.go:656] Stopping watch factory\\\\nI0930 16:12:24.709331 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 16:12:24.709334 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 16:12:24.709351 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 16:12:24.709344 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 16:12:24.709367 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:24.709376 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 16:12:24.709396 6077 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.319108 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.319174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.319194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.319229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.319247 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.346153 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.368893 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.383323 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.399155 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:26Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.422656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.422701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.422712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.422730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.422743 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.525351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.525406 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.525426 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.525448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.525464 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.628372 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.628421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.628432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.628450 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.628462 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.731244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.731360 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.731384 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.731417 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.731438 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.834800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.834850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.834860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.834879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.834922 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.938505 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.938563 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.938575 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.938598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:26 crc kubenswrapper[4796]: I0930 16:12:26.938618 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:26Z","lastTransitionTime":"2025-09-30T16:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.034615 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/1.log" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.035893 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/0.log" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.039967 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0" exitCode=1 Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.040088 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.040184 4796 scope.go:117] "RemoveContainer" containerID="302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.041089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.041139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.041155 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.041178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.041196 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.041327 4796 scope.go:117] "RemoveContainer" containerID="10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0" Sep 30 16:12:27 crc kubenswrapper[4796]: E0930 16:12:27.041606 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.077360 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.094648 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.114640 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.127616 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.143720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.143777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.143789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.143811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.143829 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.144312 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.160508 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.171492 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.185422 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.200567 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.213274 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.232515 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.248553 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.248606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.248620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.248648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.248663 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.252932 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.276040 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.290887 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.313413 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://302e0aa9ca65b999fcd11f1be4c76608157542eb284a2113cdccf653fc5c7bf5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:24.709145 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:24.709163 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 16:12:24.709179 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 16:12:24.709194 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 16:12:24.709237 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 16:12:24.709255 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:24.709260 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 16:12:24.709305 6077 factory.go:656] Stopping watch factory\\\\nI0930 16:12:24.709331 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 16:12:24.709334 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 16:12:24.709351 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 16:12:24.709344 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 16:12:24.709367 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:24.709376 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 16:12:24.709396 6077 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:27Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.351244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.351299 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.351318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.351348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.351366 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.458490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.458581 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.458603 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.458631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.458653 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.561568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.561924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.562094 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.562228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.562320 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.665629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.665696 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.665709 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.665732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.665747 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.732454 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.732587 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.732454 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:27 crc kubenswrapper[4796]: E0930 16:12:27.732642 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:27 crc kubenswrapper[4796]: E0930 16:12:27.732767 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:27 crc kubenswrapper[4796]: E0930 16:12:27.732931 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.769174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.769239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.769259 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.769285 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.769303 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.873324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.873429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.873501 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.873565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.873589 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.976657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.976708 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.976720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.976744 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:27 crc kubenswrapper[4796]: I0930 16:12:27.976757 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:27Z","lastTransitionTime":"2025-09-30T16:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.045336 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/1.log" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.051695 4796 scope.go:117] "RemoveContainer" containerID="10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0" Sep 30 16:12:28 crc kubenswrapper[4796]: E0930 16:12:28.052156 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.079868 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.080517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.080574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.080598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.080623 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.080640 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.097750 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.114354 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.135867 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.151145 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.180458 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.184076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.184137 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.184151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.184171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.184186 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.199628 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.216898 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.233399 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.253307 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.283623 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.286967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.287042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.287054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.287085 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.287100 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.304062 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.323136 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.345117 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.368064 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.391046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.391100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.391141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.391194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.391212 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.458744 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7"] Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.459492 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.462637 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.462895 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.485027 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.494420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.494481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.494494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.494514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.494529 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.504941 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.541850 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.557692 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.575517 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.585833 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t7tc\" (UniqueName: \"kubernetes.io/projected/75a4b175-f2f2-4770-8463-7dba302f15c5-kube-api-access-8t7tc\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.585976 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75a4b175-f2f2-4770-8463-7dba302f15c5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.586071 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75a4b175-f2f2-4770-8463-7dba302f15c5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.586104 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75a4b175-f2f2-4770-8463-7dba302f15c5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.592403 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.597305 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.597359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.597373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.597393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.597406 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.611272 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.650948 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.676511 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.686837 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75a4b175-f2f2-4770-8463-7dba302f15c5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.686904 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75a4b175-f2f2-4770-8463-7dba302f15c5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.686935 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75a4b175-f2f2-4770-8463-7dba302f15c5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.686960 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t7tc\" (UniqueName: \"kubernetes.io/projected/75a4b175-f2f2-4770-8463-7dba302f15c5-kube-api-access-8t7tc\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.687558 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75a4b175-f2f2-4770-8463-7dba302f15c5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.687915 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75a4b175-f2f2-4770-8463-7dba302f15c5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.696540 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75a4b175-f2f2-4770-8463-7dba302f15c5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.700811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.700850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.700861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.700877 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.700890 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.712489 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.723571 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t7tc\" (UniqueName: \"kubernetes.io/projected/75a4b175-f2f2-4770-8463-7dba302f15c5-kube-api-access-8t7tc\") pod \"ovnkube-control-plane-749d76644c-bmzf7\" (UID: \"75a4b175-f2f2-4770-8463-7dba302f15c5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.728312 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.741608 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.759669 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.771431 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.782876 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.788441 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.802670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.802710 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.802719 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.802755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.802769 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.805771 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.821630 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.836461 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.847817 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.864137 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.878420 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.893329 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.906041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.906089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.906103 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.906122 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.906138 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:28Z","lastTransitionTime":"2025-09-30T16:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.909720 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.924928 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.940295 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.962294 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:28 crc kubenswrapper[4796]: I0930 16:12:28.982595 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.001105 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:28Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.009330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.009369 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.009385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.009407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.009423 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.016556 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.031610 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.054357 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.056614 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" event={"ID":"75a4b175-f2f2-4770-8463-7dba302f15c5","Type":"ContainerStarted","Data":"92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.056664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" event={"ID":"75a4b175-f2f2-4770-8463-7dba302f15c5","Type":"ContainerStarted","Data":"700eb8ad4ee8e337a6a2e6f9abba547c403835f277efbaa8105e55077be10351"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.072944 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.111434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.111463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.111472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.111487 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.111496 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.191041 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-tl9z9"] Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.191760 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.191857 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.214705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.214756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.214769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.214789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.214802 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.215919 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.236095 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.252239 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.275299 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.290759 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.294179 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g5mv\" (UniqueName: \"kubernetes.io/projected/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-kube-api-access-8g5mv\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.294272 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.304874 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.316422 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.317497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.317525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.317534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.317549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.317560 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.327013 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.337125 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.348278 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.357112 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.371054 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.382097 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.395710 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g5mv\" (UniqueName: \"kubernetes.io/projected/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-kube-api-access-8g5mv\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.395776 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.395860 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.395911 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:29.895897546 +0000 UTC m=+41.909176073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.397845 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.411703 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.420491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.420820 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.420830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.420846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.420857 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.421036 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g5mv\" (UniqueName: \"kubernetes.io/projected/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-kube-api-access-8g5mv\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.430591 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.459071 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:29Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.524779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.524860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.524878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.524905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.524931 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.628353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.628391 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.628402 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.628421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.628433 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.731410 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.731448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.731459 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.731475 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.731485 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.732406 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.732479 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.732411 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.732602 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.732743 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.732873 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.835100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.835180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.835199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.835226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.835244 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.901820 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.901947 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:29 crc kubenswrapper[4796]: E0930 16:12:29.902032 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:30.902015936 +0000 UTC m=+42.915294463 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.939248 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.939296 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.939305 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.939321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:29 crc kubenswrapper[4796]: I0930 16:12:29.939332 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:29Z","lastTransitionTime":"2025-09-30T16:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.042166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.042244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.042277 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.042305 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.042326 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.061627 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" event={"ID":"75a4b175-f2f2-4770-8463-7dba302f15c5","Type":"ContainerStarted","Data":"119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.099513 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.121733 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.139967 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.144520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.144573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.144590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.144612 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.144659 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.152860 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.166127 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.181367 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.194765 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.211927 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.229164 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.244810 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.247117 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.247164 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.247175 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.247193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.247206 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.269696 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.285032 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.306972 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.326939 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.345813 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.350719 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.350789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.350811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.350841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.350860 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.377277 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.393898 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:30Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.454508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.454579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.454600 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.454626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.454645 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.557327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.557386 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.557398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.557442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.557465 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.659388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.659428 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.659439 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.659455 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.659467 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.732640 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:30 crc kubenswrapper[4796]: E0930 16:12:30.733309 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.762100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.762159 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.762180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.762206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.762225 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.865118 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.865173 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.865190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.865213 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.865230 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.913590 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:30 crc kubenswrapper[4796]: E0930 16:12:30.913732 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:30 crc kubenswrapper[4796]: E0930 16:12:30.913787 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:32.913773141 +0000 UTC m=+44.927051668 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.968606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.968672 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.968689 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.968716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:30 crc kubenswrapper[4796]: I0930 16:12:30.968734 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:30Z","lastTransitionTime":"2025-09-30T16:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.072052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.072133 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.072166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.072194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.072214 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.175698 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.175741 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.175752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.175770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.175783 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.279075 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.279147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.279166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.279197 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.279217 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.382085 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.382128 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.382139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.382191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.382205 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.485290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.485370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.485397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.485432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.485458 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.589163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.589241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.589265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.589299 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.589322 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.692278 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.692358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.692379 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.692408 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.692432 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.732957 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.733047 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:31 crc kubenswrapper[4796]: E0930 16:12:31.733179 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.733259 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:31 crc kubenswrapper[4796]: E0930 16:12:31.733339 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:31 crc kubenswrapper[4796]: E0930 16:12:31.733457 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.795026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.795082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.795095 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.795115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.795128 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.898421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.898475 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.898493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.898517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:31 crc kubenswrapper[4796]: I0930 16:12:31.898534 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:31Z","lastTransitionTime":"2025-09-30T16:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.002346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.002415 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.002435 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.002459 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.002477 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.104762 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.104822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.104842 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.104865 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.104883 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.208129 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.208211 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.208283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.208313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.208362 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.311635 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.311698 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.311717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.311740 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.311759 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.415054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.415121 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.415144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.415175 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.415198 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.517752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.517819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.517842 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.517874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.517897 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.620731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.620803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.620834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.620868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.620894 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.724309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.724401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.724421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.724449 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.724469 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.732282 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:32 crc kubenswrapper[4796]: E0930 16:12:32.732470 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.827156 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.827221 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.827239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.827263 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.827282 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.930591 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.930663 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.930681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.930703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.930729 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:32Z","lastTransitionTime":"2025-09-30T16:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:32 crc kubenswrapper[4796]: I0930 16:12:32.934249 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:32 crc kubenswrapper[4796]: E0930 16:12:32.934363 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:32 crc kubenswrapper[4796]: E0930 16:12:32.934402 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:36.93438869 +0000 UTC m=+48.947667217 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.033790 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.033862 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.033881 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.033906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.033925 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.137005 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.137074 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.137095 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.137124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.137145 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.240490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.240556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.240573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.240595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.240612 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.343145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.343211 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.343234 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.343266 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.343292 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.446114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.446158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.446167 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.446181 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.446191 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.549244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.549322 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.549348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.549381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.549407 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.652115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.652202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.652229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.652265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.652288 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.733213 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.733289 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.733364 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:33 crc kubenswrapper[4796]: E0930 16:12:33.733469 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:33 crc kubenswrapper[4796]: E0930 16:12:33.734121 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:33 crc kubenswrapper[4796]: E0930 16:12:33.734474 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.755443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.755497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.755510 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.755530 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.755543 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.858611 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.858662 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.858676 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.858694 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.858708 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.961680 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.961721 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.961732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.961748 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:33 crc kubenswrapper[4796]: I0930 16:12:33.961758 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:33Z","lastTransitionTime":"2025-09-30T16:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.064765 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.064808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.064820 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.064836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.064847 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.167098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.167161 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.167177 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.167202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.167221 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.208239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.208283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.208291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.208308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.208319 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.224254 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:34Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.228588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.228621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.228632 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.228649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.228660 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.245619 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:34Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.250464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.250519 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.250532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.250554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.250571 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.267083 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:34Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.271444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.271493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.271505 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.271528 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.271540 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.289006 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:34Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.293434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.293640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.293812 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.293967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.294154 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.308863 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:34Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.309031 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.311099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.311138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.311150 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.311171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.311185 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.414136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.414178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.414189 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.414207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.414221 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.516765 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.516822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.516833 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.516852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.516864 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.620036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.620083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.620091 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.620108 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.620121 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.687270 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.688671 4796 scope.go:117] "RemoveContainer" containerID="10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0" Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.688942 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.723495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.723559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.723579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.723605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.723622 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.733066 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:34 crc kubenswrapper[4796]: E0930 16:12:34.733235 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.826794 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.826861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.826882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.826907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.826924 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.929880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.929917 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.929926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.929941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:34 crc kubenswrapper[4796]: I0930 16:12:34.929952 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:34Z","lastTransitionTime":"2025-09-30T16:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.032873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.032932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.032952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.033027 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.033048 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.136144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.136200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.136212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.136231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.136242 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.239152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.239224 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.239268 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.239295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.239333 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.342389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.342455 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.342473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.342496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.342513 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.446100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.446193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.446215 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.446239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.446259 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.548927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.549029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.549056 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.549094 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.549120 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.655055 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.655126 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.655152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.655182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.655204 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.732790 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.732897 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.733033 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:35 crc kubenswrapper[4796]: E0930 16:12:35.733098 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:35 crc kubenswrapper[4796]: E0930 16:12:35.733294 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:35 crc kubenswrapper[4796]: E0930 16:12:35.733432 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.758137 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.758207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.758226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.758253 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.758271 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.861329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.861415 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.861440 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.861474 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.861497 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.964301 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.964368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.964387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.964411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:35 crc kubenswrapper[4796]: I0930 16:12:35.964428 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:35Z","lastTransitionTime":"2025-09-30T16:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.067750 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.067820 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.067843 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.067875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.067896 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.171226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.171311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.171363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.171399 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.171417 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.274285 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.274332 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.274343 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.274364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.274377 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.376954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.377077 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.377100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.377128 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.377148 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.480596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.480649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.480666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.480690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.480706 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.584424 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.584506 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.584531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.584558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.584578 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.687175 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.687246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.687269 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.687301 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.687334 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.733287 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:36 crc kubenswrapper[4796]: E0930 16:12:36.733459 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.789889 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.789940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.789958 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.790010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.790027 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.892661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.892743 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.892766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.892797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.892822 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.975190 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.976085 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:36 crc kubenswrapper[4796]: E0930 16:12:36.976288 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:36 crc kubenswrapper[4796]: E0930 16:12:36.976414 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:12:44.976386398 +0000 UTC m=+56.989664965 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.985459 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.992462 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:36Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.996072 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.996130 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.996150 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.996177 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:36 crc kubenswrapper[4796]: I0930 16:12:36.996197 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:36Z","lastTransitionTime":"2025-09-30T16:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.008949 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.035198 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.050102 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.063399 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.077102 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.093534 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.098307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.098377 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.098390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.098415 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.098427 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.106189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.124612 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.141363 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.157570 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.174745 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.191718 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.201468 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.201506 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.201515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.201531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.201543 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.210492 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.242496 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.265855 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.286638 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:37Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.304835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.304927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.304939 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.304965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.304999 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.408323 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.408374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.408387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.408405 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.408418 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.511195 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.511276 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.511297 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.511323 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.511341 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.615021 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.615090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.615107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.615132 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.615150 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.717904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.718031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.718055 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.718082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.718105 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.732546 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.732639 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:37 crc kubenswrapper[4796]: E0930 16:12:37.732720 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.732657 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:37 crc kubenswrapper[4796]: E0930 16:12:37.732849 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:37 crc kubenswrapper[4796]: E0930 16:12:37.733046 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.821638 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.821687 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.821699 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.821719 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.821732 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.925013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.925076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.925098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.925127 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:37 crc kubenswrapper[4796]: I0930 16:12:37.925151 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:37Z","lastTransitionTime":"2025-09-30T16:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.028138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.028212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.028232 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.028256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.028273 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.130398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.130476 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.130499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.130529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.130552 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.233561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.233633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.233655 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.233688 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.233711 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.337027 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.337087 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.337106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.337132 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.337149 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.440499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.440622 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.440653 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.440685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.440709 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.544461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.544522 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.544540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.544569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.544591 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.647602 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.647646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.647657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.647676 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.647688 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.733515 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:38 crc kubenswrapper[4796]: E0930 16:12:38.734449 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.749946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.750048 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.750069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.750096 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.750114 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.760869 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.778503 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.796116 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.811957 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.835614 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.850556 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.852083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.852116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.852126 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.852143 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.852157 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.865189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.877639 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.895641 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.942060 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.954459 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.954519 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.954537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.954561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.954579 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:38Z","lastTransitionTime":"2025-09-30T16:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.962379 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:38 crc kubenswrapper[4796]: I0930 16:12:38.988393 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:38Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.003280 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:39Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.018237 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:39Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.042379 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:39Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.057248 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:39Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.057450 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.057473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.057484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.057500 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.057509 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.070142 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:39Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.083224 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:39Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.159922 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.160022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.160050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.160084 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.160102 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.263486 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.263556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.263580 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.263610 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.263634 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.366157 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.366215 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.366235 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.366259 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.366276 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.468876 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.468954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.469018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.469051 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.469074 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.506545 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.506729 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.506744 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:11.506718629 +0000 UTC m=+83.519997166 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.506799 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.506885 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.506918 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.507021 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:13:11.506970027 +0000 UTC m=+83.520248564 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.507047 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:13:11.507036189 +0000 UTC m=+83.520314866 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.572316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.572387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.572404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.572430 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.572450 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.607766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.607889 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608139 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608185 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608191 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608259 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608290 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608209 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608393 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:13:11.608357369 +0000 UTC m=+83.621635956 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.608504 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:13:11.608471323 +0000 UTC m=+83.621749930 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.675613 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.675680 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.675700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.675724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.675742 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.732715 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.732757 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.732840 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.732732 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.733014 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:39 crc kubenswrapper[4796]: E0930 16:12:39.733110 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.778963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.779021 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.779032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.779047 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.779060 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.882237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.882308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.882331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.882359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.882379 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.985450 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.985508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.985531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.985559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:39 crc kubenswrapper[4796]: I0930 16:12:39.985577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:39Z","lastTransitionTime":"2025-09-30T16:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.089229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.089283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.089303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.089326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.089343 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.192903 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.193357 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.193444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.193522 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.193628 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.297138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.297198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.297216 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.297241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.297261 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.400012 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.400100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.400124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.400158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.400185 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.503708 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.503772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.503791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.503818 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.503836 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.607345 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.607408 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.607425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.607447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.607464 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.710323 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.710363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.710376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.710394 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.710405 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.733505 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:40 crc kubenswrapper[4796]: E0930 16:12:40.733747 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.812759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.812831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.812855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.812884 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.812906 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.916068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.916123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.916140 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.916163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:40 crc kubenswrapper[4796]: I0930 16:12:40.916180 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:40Z","lastTransitionTime":"2025-09-30T16:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.018090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.018152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.018169 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.018198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.018215 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.120992 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.121051 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.121114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.121139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.121151 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.224151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.224224 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.224242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.224269 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.224290 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.327351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.327409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.327431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.327456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.327480 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.430525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.430595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.430620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.430648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.430676 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.532749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.532828 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.532850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.532878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.532900 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.635766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.635838 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.635860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.635889 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.635907 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.732643 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.732748 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.732658 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:41 crc kubenswrapper[4796]: E0930 16:12:41.732862 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:41 crc kubenswrapper[4796]: E0930 16:12:41.732945 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:41 crc kubenswrapper[4796]: E0930 16:12:41.733129 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.738773 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.738843 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.738868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.738899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.738923 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.842805 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.842863 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.842880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.842906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.842925 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.950555 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.950608 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.950626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.950650 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:41 crc kubenswrapper[4796]: I0930 16:12:41.950670 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:41Z","lastTransitionTime":"2025-09-30T16:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.053004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.053062 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.053074 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.053093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.053106 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.156571 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.156712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.156727 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.156749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.156761 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.260148 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.260229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.260253 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.260286 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.260312 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.363697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.363756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.363766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.363781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.363791 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.467425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.467473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.467484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.467503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.467515 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.570265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.570307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.570318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.570337 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.570348 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.672620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.672679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.672697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.672718 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.672733 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.732887 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:42 crc kubenswrapper[4796]: E0930 16:12:42.733159 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.774891 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.774961 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.775001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.775018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.775030 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.877972 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.878026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.878036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.878052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.878063 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.981292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.981350 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.981364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.981385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:42 crc kubenswrapper[4796]: I0930 16:12:42.981401 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:42Z","lastTransitionTime":"2025-09-30T16:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.084786 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.084862 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.084886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.084916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.084938 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.187692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.187723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.187732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.187745 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.187754 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.290257 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.290294 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.290307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.290325 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.290334 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.393847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.393905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.393918 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.393945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.393958 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.496885 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.496967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.497028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.497063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.497090 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.599845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.600947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.601019 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.601046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.601064 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.703434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.703502 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.703520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.703543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.703563 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.733305 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.733297 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:43 crc kubenswrapper[4796]: E0930 16:12:43.733542 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:43 crc kubenswrapper[4796]: E0930 16:12:43.733578 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.733329 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:43 crc kubenswrapper[4796]: E0930 16:12:43.733651 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.806018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.806088 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.806104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.806124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.806138 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.909005 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.909055 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.909067 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.909087 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:43 crc kubenswrapper[4796]: I0930 16:12:43.909100 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:43Z","lastTransitionTime":"2025-09-30T16:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.012275 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.012353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.012372 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.012404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.012456 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.115839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.115892 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.115928 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.115946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.115960 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.219324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.219410 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.219438 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.219472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.219494 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.325812 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.325932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.326036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.326067 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.326091 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.373697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.373774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.373799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.373831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.373855 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.391535 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:44Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.396625 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.396687 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.396709 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.396739 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.396763 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.412886 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:44Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.418118 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.418182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.418207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.418235 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.418292 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.436952 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:44Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.442029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.442111 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.442138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.442164 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.442181 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.464179 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:44Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.471537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.471626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.471645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.471670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.471688 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.492168 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:44Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.492328 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.494801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.494861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.494878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.494905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.494923 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.598249 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.598304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.598316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.598342 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.598357 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.700602 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.700639 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.700650 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.700667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.700679 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.733305 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:44 crc kubenswrapper[4796]: E0930 16:12:44.733582 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.803226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.803320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.803336 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.803353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.803365 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.906956 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.907032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.907049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.907070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:44 crc kubenswrapper[4796]: I0930 16:12:44.907083 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:44Z","lastTransitionTime":"2025-09-30T16:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.010399 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.010464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.010482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.010507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.010526 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.070350 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:45 crc kubenswrapper[4796]: E0930 16:12:45.070584 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:45 crc kubenswrapper[4796]: E0930 16:12:45.070932 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:13:01.070910158 +0000 UTC m=+73.084188685 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.113489 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.113552 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.113577 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.113605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.113629 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.217345 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.217410 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.217433 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.217462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.217487 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.319841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.319897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.319909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.319931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.319943 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.422540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.422600 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.422612 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.422627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.422637 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.525107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.525183 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.525193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.525208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.525218 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.627954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.628022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.628039 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.628055 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.628066 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.731020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.731070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.731082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.731099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.731109 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.732533 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.732579 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.732640 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:45 crc kubenswrapper[4796]: E0930 16:12:45.732815 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:45 crc kubenswrapper[4796]: E0930 16:12:45.732935 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:45 crc kubenswrapper[4796]: E0930 16:12:45.733152 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.834237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.834284 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.834297 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.834320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.834336 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.937550 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.937608 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.937619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.937640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:45 crc kubenswrapper[4796]: I0930 16:12:45.937656 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:45Z","lastTransitionTime":"2025-09-30T16:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.041405 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.041873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.042054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.042203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.042375 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.145309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.145368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.145379 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.145404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.145418 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.248866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.248924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.248936 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.248959 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.248973 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.352692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.352781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.352799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.352822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.352841 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.456328 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.456375 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.456386 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.456404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.456416 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.559295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.559355 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.559372 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.559399 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.559417 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.662659 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.662721 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.662742 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.662767 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.662786 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.733024 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:46 crc kubenswrapper[4796]: E0930 16:12:46.733224 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.734281 4796 scope.go:117] "RemoveContainer" containerID="10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.767042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.767897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.768060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.768170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.768267 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.870884 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.870932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.870950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.870999 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.871016 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.973855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.973901 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.973916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.973938 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:46 crc kubenswrapper[4796]: I0930 16:12:46.973953 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:46Z","lastTransitionTime":"2025-09-30T16:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.077484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.077531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.077543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.077564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.077579 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.122379 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/1.log" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.125734 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.126254 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.149672 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.181068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.181123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.181137 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.181160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.181175 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.182082 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.199880 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.218034 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.229846 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.250492 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.264139 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.284684 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.284734 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.284748 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.284769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.284786 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.286308 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.303399 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.322087 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.338487 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.354552 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.369503 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.383894 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.388109 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.388145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.388153 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.388168 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.388179 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.396325 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.411261 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.422210 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.435691 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:47Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.490674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.490713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.490722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.490735 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.490746 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.593970 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.594303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.594387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.594484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.594569 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.697291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.697340 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.697355 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.697373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.697386 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.732423 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.732480 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:47 crc kubenswrapper[4796]: E0930 16:12:47.732577 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.732649 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:47 crc kubenswrapper[4796]: E0930 16:12:47.732798 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:47 crc kubenswrapper[4796]: E0930 16:12:47.732657 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.801376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.801445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.801469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.801499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.801522 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.904971 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.905098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.905124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.905153 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:47 crc kubenswrapper[4796]: I0930 16:12:47.905175 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:47Z","lastTransitionTime":"2025-09-30T16:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.008031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.008089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.008100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.008118 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.008130 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.111540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.111609 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.111633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.111661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.111684 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.132022 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/2.log" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.133444 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/1.log" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.137423 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" exitCode=1 Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.137479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.137528 4796 scope.go:117] "RemoveContainer" containerID="10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.138783 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:12:48 crc kubenswrapper[4796]: E0930 16:12:48.139145 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.161433 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.179389 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.197592 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.214450 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.214514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.214527 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.214547 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.214558 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.222908 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.240006 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.259684 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.278838 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.293617 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.307144 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.317113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.317191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.317218 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.317249 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.317282 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.331036 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.348419 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.376142 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.391522 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.409207 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.419451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.419487 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.419498 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.419516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.419527 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.421834 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.433239 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.447810 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.461492 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.521263 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.521304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.521312 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.521326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.521336 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.624279 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.624339 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.624354 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.624383 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.624399 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.726780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.726823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.726832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.726846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.726856 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.732227 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:48 crc kubenswrapper[4796]: E0930 16:12:48.732350 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.773190 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.790335 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.806359 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.828283 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.829160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.829221 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.829240 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.829621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.829651 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.851257 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.867226 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.887301 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.904087 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.920298 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.933391 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.933456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.933477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.933503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.933521 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:48Z","lastTransitionTime":"2025-09-30T16:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.942137 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.958946 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.974776 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:48 crc kubenswrapper[4796]: I0930 16:12:48.994281 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:48Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.011903 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.033009 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.035818 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.035867 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.035883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.035911 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.035947 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.051260 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.065959 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.089208 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10fae5a2fba553efba13342eb906691195c78839977c41805a1309478c56c2e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:26Z\\\",\\\"message\\\":\\\"11 6247 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488473 6247 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:26.488566 6247 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488764 6247 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488820 6247 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.488877 6247 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:26.489218 6247 factory.go:656] Stopping watch factory\\\\nI0930 16:12:26.520786 6247 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:26.520838 6247 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:26.520946 6247 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:26.521022 6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:26.521109 6247 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.138947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.139028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.139039 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.139057 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.139092 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.142538 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/2.log" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.148656 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:12:49 crc kubenswrapper[4796]: E0930 16:12:49.149209 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.163719 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.183466 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.197510 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.214866 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.229340 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.241678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.241731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.241745 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.241766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.241782 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.245677 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.261796 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.277052 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.294100 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.318420 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.333184 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.344765 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.344849 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.344869 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.344894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.344915 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.355854 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.378471 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.397342 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.414328 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.426211 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.439725 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.448195 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.448494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.448715 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.448920 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.449133 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.453261 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:49Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.551518 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.551909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.552016 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.552111 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.552173 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.655351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.655427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.655447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.655473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.655492 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.732255 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.732284 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.732350 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:49 crc kubenswrapper[4796]: E0930 16:12:49.732458 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:49 crc kubenswrapper[4796]: E0930 16:12:49.732576 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:49 crc kubenswrapper[4796]: E0930 16:12:49.732675 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.758550 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.758623 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.758649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.758680 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.758704 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.861701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.861780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.861799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.861829 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.861846 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.964735 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.964796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.964808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.964824 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:49 crc kubenswrapper[4796]: I0930 16:12:49.964835 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:49Z","lastTransitionTime":"2025-09-30T16:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.068463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.068523 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.068549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.068580 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.068602 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.173080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.173149 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.173171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.173202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.173225 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.276443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.276508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.276526 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.276556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.276577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.380373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.380454 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.380476 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.380510 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.380535 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.483847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.483882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.483891 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.483906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.483915 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.586588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.586643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.586662 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.586688 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.586706 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.690059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.690109 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.690120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.690141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.690155 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.732779 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:50 crc kubenswrapper[4796]: E0930 16:12:50.732971 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.792782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.792816 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.792829 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.792845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.792858 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.896184 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.896246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.896259 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.896280 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.896297 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.999256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.999321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.999339 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.999362 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:50 crc kubenswrapper[4796]: I0930 16:12:50.999382 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:50Z","lastTransitionTime":"2025-09-30T16:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.102082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.102157 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.102194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.102228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.102248 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.205145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.205205 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.205222 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.205245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.205261 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.307965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.308086 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.308154 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.308174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.308185 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.411172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.411221 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.411249 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.411265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.411275 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.517826 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.517925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.517942 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.517964 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.517996 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.620539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.620629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.620648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.620676 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.620695 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.723513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.723576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.723593 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.723622 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.723639 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.733083 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.733136 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.733137 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:51 crc kubenswrapper[4796]: E0930 16:12:51.733280 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:51 crc kubenswrapper[4796]: E0930 16:12:51.733495 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:51 crc kubenswrapper[4796]: E0930 16:12:51.733620 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.825625 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.825675 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.825689 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.825710 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.825724 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.928245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.928292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.928303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.928321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:51 crc kubenswrapper[4796]: I0930 16:12:51.928330 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:51Z","lastTransitionTime":"2025-09-30T16:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.030901 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.031211 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.031716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.031784 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.032156 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.134815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.134866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.134883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.134906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.134924 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.237806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.237865 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.237875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.237890 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.237920 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.341394 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.341481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.341513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.341543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.341567 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.445713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.445772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.445790 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.445815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.445832 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.548726 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.548797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.548819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.548845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.548865 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.651350 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.651389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.651397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.651413 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.651422 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.732288 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:52 crc kubenswrapper[4796]: E0930 16:12:52.732507 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.753944 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.754013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.754030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.754046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.754058 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.856181 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.856228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.856242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.856262 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.856274 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.958667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.958717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.958730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.958749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:52 crc kubenswrapper[4796]: I0930 16:12:52.958762 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:52Z","lastTransitionTime":"2025-09-30T16:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.061330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.061369 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.061381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.061396 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.061407 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.163262 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.163335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.163348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.163392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.163405 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.265872 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.265937 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.265962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.266032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.266059 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.368692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.368737 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.368749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.368766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.368780 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.471793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.471844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.471855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.471871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.471882 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.575058 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.575096 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.575107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.575120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.575129 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.677862 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.677926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.677952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.677974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.678012 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.732742 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.732786 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.732838 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:53 crc kubenswrapper[4796]: E0930 16:12:53.732909 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:53 crc kubenswrapper[4796]: E0930 16:12:53.733106 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:53 crc kubenswrapper[4796]: E0930 16:12:53.733245 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.780227 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.780283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.780294 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.780313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.780322 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.882884 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.882924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.882932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.882945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.882955 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.985370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.985436 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.985445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.985460 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:53 crc kubenswrapper[4796]: I0930 16:12:53.985482 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:53Z","lastTransitionTime":"2025-09-30T16:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.088720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.088762 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.088770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.088785 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.088796 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.191852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.191915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.191927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.191948 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.191961 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.295202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.295255 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.295271 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.295290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.295303 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.397927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.398059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.398072 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.398098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.398114 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.501537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.501598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.501612 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.501627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.501638 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.604118 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.604182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.604200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.604225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.604244 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.706671 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.707173 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.707185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.707200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.707212 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.733161 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.733343 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.750257 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.810258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.810304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.810313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.810328 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.810337 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.817812 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.817844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.817852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.817888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.817897 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.828944 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:54Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.833009 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.833250 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.833269 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.833288 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.833301 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.847788 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:54Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.852279 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.852308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.852317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.852330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.852340 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.868595 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:54Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.872777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.872830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.872849 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.872874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.872891 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.884678 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:54Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.888187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.888227 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.888240 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.888256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.888269 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.898859 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:54Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:54 crc kubenswrapper[4796]: E0930 16:12:54.899073 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.912508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.912560 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.912570 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.912584 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:54 crc kubenswrapper[4796]: I0930 16:12:54.912593 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:54Z","lastTransitionTime":"2025-09-30T16:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.015943 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.016017 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.016040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.016068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.016088 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.119598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.119646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.119657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.119674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.119686 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.222400 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.222482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.222506 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.222528 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.222545 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.324962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.325018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.325028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.325044 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.325056 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.427508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.427560 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.427568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.427583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.427594 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.530908 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.530965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.530997 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.531015 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.531027 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.633419 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.633456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.633466 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.633482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.633496 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.732252 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.732279 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.732358 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:55 crc kubenswrapper[4796]: E0930 16:12:55.732399 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:55 crc kubenswrapper[4796]: E0930 16:12:55.732497 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:55 crc kubenswrapper[4796]: E0930 16:12:55.732741 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.736367 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.736448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.736461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.736477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.736486 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.839119 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.839191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.839205 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.839231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.839250 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.942452 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.942516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.942534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.942559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:55 crc kubenswrapper[4796]: I0930 16:12:55.942577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:55Z","lastTransitionTime":"2025-09-30T16:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.046204 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.046272 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.046290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.046315 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.046332 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.149188 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.149257 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.149275 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.149301 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.149322 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.252419 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.252496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.252505 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.252525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.252537 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.355390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.355457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.355467 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.355491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.355507 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.458103 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.458153 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.458166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.458185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.458197 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.561153 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.561207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.561217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.561237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.561249 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.663714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.663753 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.663765 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.663782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.663794 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.732866 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:56 crc kubenswrapper[4796]: E0930 16:12:56.733060 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.766772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.766824 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.766834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.766851 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.766863 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.870329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.870393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.870414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.870448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.870470 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.973020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.973329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.973477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.973633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:56 crc kubenswrapper[4796]: I0930 16:12:56.973799 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:56Z","lastTransitionTime":"2025-09-30T16:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.076927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.077001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.077021 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.077042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.077054 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.179160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.179203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.179213 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.179227 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.179236 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.282268 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.282596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.282696 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.282793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.282874 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.385312 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.385652 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.385836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.385974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.386176 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.489592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.489653 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.489671 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.489695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.489714 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.592714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.592762 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.592775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.592792 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.592805 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.696053 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.696089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.696097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.696112 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.696121 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.732720 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.732789 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:57 crc kubenswrapper[4796]: E0930 16:12:57.733363 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:57 crc kubenswrapper[4796]: E0930 16:12:57.733185 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.732840 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:57 crc kubenswrapper[4796]: E0930 16:12:57.733473 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.798594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.798643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.798656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.798673 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.798683 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.901320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.902371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.902492 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.902576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:57 crc kubenswrapper[4796]: I0930 16:12:57.902661 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:57Z","lastTransitionTime":"2025-09-30T16:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.005600 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.006105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.006278 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.006486 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.006666 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.110191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.110255 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.110274 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.110296 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.110315 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.213958 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.214488 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.214629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.214820 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.214976 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.318034 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.318090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.318102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.318126 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.318145 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.423010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.423076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.423089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.423113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.423131 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.525930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.526036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.526049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.526073 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.526089 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.629113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.629179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.629196 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.629222 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.629239 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.732293 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.732358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.732376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.732401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.732418 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.732726 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:12:58 crc kubenswrapper[4796]: E0930 16:12:58.732855 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.758806 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.777503 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.790842 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.813908 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.824860 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.834144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.834182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.834191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.834208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.834221 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.848842 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.865115 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.883369 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.900600 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.917081 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.929771 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"080efbaf-9ef0-4538-aab9-624add222ac2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48af73cc5e378a96747a37927f934fb8f6591ff48a799d4cde3164a0a82053b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.936690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.936738 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.936756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.936781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.936799 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:58Z","lastTransitionTime":"2025-09-30T16:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.952162 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.968738 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.980518 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:58 crc kubenswrapper[4796]: I0930 16:12:58.992959 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:58Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.004765 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:59Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.022507 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:59Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.039552 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:59Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.039831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.039871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.039886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.039909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.039926 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.058767 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:12:59Z is after 2025-08-24T17:21:41Z" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.142805 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.142866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.142884 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.142907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.142926 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.245520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.245566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.245581 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.245602 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.245619 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.348307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.348377 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.348395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.348423 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.348447 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.450684 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.450723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.450732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.450746 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.450756 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.553097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.553184 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.553208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.553238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.553262 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.655966 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.656058 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.656077 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.656100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.656120 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.732542 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:12:59 crc kubenswrapper[4796]: E0930 16:12:59.732717 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.732850 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:12:59 crc kubenswrapper[4796]: E0930 16:12:59.733085 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.733354 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:12:59 crc kubenswrapper[4796]: E0930 16:12:59.733473 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.759486 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.759543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.759560 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.759583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.759603 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.862562 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.862628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.862644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.862667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.862684 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.965639 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.965677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.965688 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.965702 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:12:59 crc kubenswrapper[4796]: I0930 16:12:59.965711 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:12:59Z","lastTransitionTime":"2025-09-30T16:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.068290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.068342 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.068362 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.068385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.068403 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.172074 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.172139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.172156 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.172180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.172200 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.275294 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.275355 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.275371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.275395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.275413 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.378495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.378549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.378559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.378574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.378583 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.481993 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.482052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.482063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.482110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.482123 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.584761 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.584925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.584945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.584972 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.585017 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.687764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.687819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.687834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.687856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.687872 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.733025 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:00 crc kubenswrapper[4796]: E0930 16:13:00.733263 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.790897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.790947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.790965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.791020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.791039 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.893755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.893821 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.893839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.893863 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.893883 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.996888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.996940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.996952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.996969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:00 crc kubenswrapper[4796]: I0930 16:13:00.996999 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:00Z","lastTransitionTime":"2025-09-30T16:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.100484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.100549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.100569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.100595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.100611 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.138518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:01 crc kubenswrapper[4796]: E0930 16:13:01.138726 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:13:01 crc kubenswrapper[4796]: E0930 16:13:01.138831 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs podName:d1683cfa-0a4a-4f40-bbbf-6675107ecbf9 nodeName:}" failed. No retries permitted until 2025-09-30 16:13:33.138804316 +0000 UTC m=+105.152082913 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs") pod "network-metrics-daemon-tl9z9" (UID: "d1683cfa-0a4a-4f40-bbbf-6675107ecbf9") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.203693 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.203772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.203794 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.203822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.203844 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.305974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.306028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.306037 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.306050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.306060 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.408650 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.408722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.408758 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.408799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.408823 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.511307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.511377 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.511395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.511418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.511435 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.613862 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.613902 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.613915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.613934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.613946 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.716032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.716081 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.716093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.716113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.716124 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.732184 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.732240 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.732319 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:01 crc kubenswrapper[4796]: E0930 16:13:01.732453 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:01 crc kubenswrapper[4796]: E0930 16:13:01.732749 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.733033 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:13:01 crc kubenswrapper[4796]: E0930 16:13:01.733062 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:01 crc kubenswrapper[4796]: E0930 16:13:01.733192 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.818093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.818131 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.818140 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.818154 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.818165 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.921348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.921414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.921431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.921457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:01 crc kubenswrapper[4796]: I0930 16:13:01.921475 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:01Z","lastTransitionTime":"2025-09-30T16:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.024775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.024832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.024855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.024879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.024894 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.127217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.127276 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.127291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.127316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.127336 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.229856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.229919 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.229943 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.229969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.230021 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.332879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.332912 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.332921 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.332934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.332946 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.435415 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.435483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.435499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.435520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.435536 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.539324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.539389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.539413 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.539443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.539465 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.643397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.643463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.643481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.643505 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.643522 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.732381 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:02 crc kubenswrapper[4796]: E0930 16:13:02.732635 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.745872 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.745906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.745914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.745925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.745935 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.849018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.849109 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.849122 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.849144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.849158 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.952839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.952931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.952943 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.952963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:02 crc kubenswrapper[4796]: I0930 16:13:02.953016 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:02Z","lastTransitionTime":"2025-09-30T16:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.056714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.056777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.056800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.056827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.056844 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.160085 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.160191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.160209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.160235 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.160252 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.196013 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hldp6_b4cf8ea3-7aa4-4dfc-8443-347fd3898f15/kube-multus/0.log" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.196097 4796 generic.go:334] "Generic (PLEG): container finished" podID="b4cf8ea3-7aa4-4dfc-8443-347fd3898f15" containerID="1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f" exitCode=1 Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.196196 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hldp6" event={"ID":"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15","Type":"ContainerDied","Data":"1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.196954 4796 scope.go:117] "RemoveContainer" containerID="1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.213293 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.235490 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.254751 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.263791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.263885 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.263914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.263962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.264021 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.274410 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.292875 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.310624 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.329933 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.347155 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.360864 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.367364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.367429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.367442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.367460 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.367472 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.382217 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.398555 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.420551 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.442082 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.468714 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.471104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.471160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.471180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.471215 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.471239 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.486049 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.502529 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.515836 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"080efbaf-9ef0-4538-aab9-624add222ac2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48af73cc5e378a96747a37927f934fb8f6591ff48a799d4cde3164a0a82053b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.534508 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.553802 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:13:02Z\\\",\\\"message\\\":\\\"2025-09-30T16:12:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d2e7a03b-5b56-4681-a4de-b8727b3d3fce\\\\n2025-09-30T16:12:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d2e7a03b-5b56-4681-a4de-b8727b3d3fce to /host/opt/cni/bin/\\\\n2025-09-30T16:12:17Z [verbose] multus-daemon started\\\\n2025-09-30T16:12:17Z [verbose] Readiness Indicator file check\\\\n2025-09-30T16:13:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:03Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.573958 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.574023 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.574035 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.574054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.574069 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.676952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.677020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.677030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.677045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.677058 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.732726 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.732780 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.732799 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:03 crc kubenswrapper[4796]: E0930 16:13:03.732965 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:03 crc kubenswrapper[4796]: E0930 16:13:03.733162 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:03 crc kubenswrapper[4796]: E0930 16:13:03.733339 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.780918 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.781004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.781023 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.781048 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.781066 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.883672 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.883729 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.883746 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.883769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.883785 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.986578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.986634 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.986649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.986669 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:03 crc kubenswrapper[4796]: I0930 16:13:03.986683 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:03Z","lastTransitionTime":"2025-09-30T16:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.089848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.089941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.089967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.090031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.090050 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.193295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.193378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.193413 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.193455 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.193481 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.201841 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hldp6_b4cf8ea3-7aa4-4dfc-8443-347fd3898f15/kube-multus/0.log" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.201923 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hldp6" event={"ID":"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15","Type":"ContainerStarted","Data":"24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.238933 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.261348 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.282841 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.296549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.296601 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.296613 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.296633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.296646 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.300193 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.314892 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.330366 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"080efbaf-9ef0-4538-aab9-624add222ac2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48af73cc5e378a96747a37927f934fb8f6591ff48a799d4cde3164a0a82053b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.348286 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.367180 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:13:02Z\\\",\\\"message\\\":\\\"2025-09-30T16:12:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d2e7a03b-5b56-4681-a4de-b8727b3d3fce\\\\n2025-09-30T16:12:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d2e7a03b-5b56-4681-a4de-b8727b3d3fce to /host/opt/cni/bin/\\\\n2025-09-30T16:12:17Z [verbose] multus-daemon started\\\\n2025-09-30T16:12:17Z [verbose] Readiness Indicator file check\\\\n2025-09-30T16:13:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.383602 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.399624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.399678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.399686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.399702 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.399712 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.400878 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.416940 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.437044 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.452945 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.475510 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.492446 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.502831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.502907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.502926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.502954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.502973 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.510043 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.528803 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.546641 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.577112 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:04Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.605887 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.605916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.605926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.605941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.605954 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.708357 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.708417 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.708433 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.708458 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.708474 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.733079 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:04 crc kubenswrapper[4796]: E0930 16:13:04.733295 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.811698 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.811759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.811777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.811802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.811820 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.915090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.915153 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.915174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.915200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:04 crc kubenswrapper[4796]: I0930 16:13:04.915218 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:04Z","lastTransitionTime":"2025-09-30T16:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.017803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.017867 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.017883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.017907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.017926 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.121273 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.121364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.121385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.121409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.121427 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.224918 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.225031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.225056 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.225084 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.225108 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.264160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.264234 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.264259 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.264290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.264311 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.285054 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:05Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.290358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.290420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.290437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.290462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.290479 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.311428 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:05Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.318904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.318974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.319040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.319068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.319087 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.341483 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:05Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.347143 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.347188 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.347208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.347231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.347249 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.368325 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:05Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.373521 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.373596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.373614 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.373640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.373660 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.395947 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:05Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.396193 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.398266 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.398331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.398357 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.398388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.398411 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.500571 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.500618 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.500630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.500647 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.500659 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.603321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.603388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.603404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.603432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.603450 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.706569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.706930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.706940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.707041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.707055 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.732333 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.732437 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.732501 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.732330 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.732721 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:05 crc kubenswrapper[4796]: E0930 16:13:05.732896 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.809916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.810010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.810030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.810056 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.810074 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.913223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.913285 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.913301 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.913324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:05 crc kubenswrapper[4796]: I0930 16:13:05.913341 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:05Z","lastTransitionTime":"2025-09-30T16:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.017353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.017448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.017480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.017511 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.017534 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.120963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.121078 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.121102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.121132 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.121157 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.224124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.224176 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.224188 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.224206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.224219 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.327387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.327450 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.327467 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.327493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.327513 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.430420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.430469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.430481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.430497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.430509 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.533685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.533764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.533776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.533800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.533815 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.636827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.636907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.636929 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.636961 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.637033 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.732863 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:06 crc kubenswrapper[4796]: E0930 16:13:06.733151 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.739745 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.739815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.739834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.739860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.739879 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.842705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.842763 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.842780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.842802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.842820 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.946555 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.946631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.946649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.946666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:06 crc kubenswrapper[4796]: I0930 16:13:06.946678 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:06Z","lastTransitionTime":"2025-09-30T16:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.050096 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.050148 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.050159 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.050209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.050224 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.153135 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.153199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.153216 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.153242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.153260 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.256390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.256458 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.256476 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.256499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.256516 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.359583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.359648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.359667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.359693 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.359712 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.462797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.462892 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.462913 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.463370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.463655 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.566809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.566864 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.566873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.566889 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.566899 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.669668 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.669713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.669724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.669742 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.669753 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.732936 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:07 crc kubenswrapper[4796]: E0930 16:13:07.733122 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.732961 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.733184 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:07 crc kubenswrapper[4796]: E0930 16:13:07.733335 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:07 crc kubenswrapper[4796]: E0930 16:13:07.733465 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.773489 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.773563 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.773587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.773616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.773635 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.877154 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.877206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.877218 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.877236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.877248 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.979704 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.979785 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.979803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.979825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:07 crc kubenswrapper[4796]: I0930 16:13:07.979843 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:07Z","lastTransitionTime":"2025-09-30T16:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.082425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.082463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.082475 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.082490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.082501 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.184748 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.184775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.184782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.184795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.184821 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.288050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.288112 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.288129 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.288152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.288170 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.391952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.392040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.392052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.392071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.392085 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.495174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.495216 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.495228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.495246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.495256 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.598616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.598677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.598701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.598726 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.598768 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.702132 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.702179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.702191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.702209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.702221 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.732910 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:08 crc kubenswrapper[4796]: E0930 16:13:08.733080 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.752236 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.785041 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bcf2108-84b5-4c62-b32b-ecfcd6153fc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2227fdabff265c8a07ccc74f836dbdb27cfe02fc96421dd9000bb082a7e97f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://921f56cb60b48fcb28df1ba034e9c08d8d3e4c2aa8ea8e772e5108b407e24d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92996c94eac0cfc2aa2019e09cc729691dcb36646eec92785c78507c7d800bdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba4703d40fd5ec2742da69e6eeca24f40bdb364b82d0f075c064d2aeffb5334e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d898d0236ee1ba245536fc477fbed504bc76ab149bd98f1a14372a30cec1a4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4944f41ae5f376fc7ad3799396297617a2391c9504628a6002909ffda196642a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://255dc58e16fdae580924c52dca8f1de4550478a261b4f6f7088dcb24daa2bafd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8ee6b62e6ea65fd1c73f5f9e5837dda053a8af87a9ac27983c68f7a88b01723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.805100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.805146 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.805164 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.805187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.805205 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.805568 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f3605ef-736b-44bd-8875-0ffd0325b758\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae0b80515c74a3cac4df8450a64f5f06adfda9c47211d20d03218253d0aa8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4eec3dc29708c1d7f4fbfffb33b7c177575f24ed18854624544a139e72a45eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a93e551adb199b3ba0ed6c37ee35bdb3ff6edcfb799b4c20d7b759c91f8589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac9f6c2381c9bcc7855025fbb37a72b04198b97ff453e5bba04ebd83963c4f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd754869f0be938977cbd995706d05e3ea01e896d545c88da17f74e4fb9a8029\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0930 16:12:07.474205 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0930 16:12:07.474232 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0930 16:12:07.474477 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0930 16:12:07.474527 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0930 16:12:07.474822 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\"\\\\nI0930 16:12:07.475159 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2259309738/tls.crt::/tmp/serving-cert-2259309738/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759248711\\\\\\\\\\\\\\\" (2025-09-30 16:11:51 +0000 UTC to 2025-10-30 16:11:52 +0000 UTC (now=2025-09-30 16:12:07.475078076 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475510 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759248712\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759248712\\\\\\\\\\\\\\\" (2025-09-30 15:11:52 +0000 UTC to 2026-09-30 15:11:52 +0000 UTC (now=2025-09-30 16:12:07.475459467 +0000 UTC))\\\\\\\"\\\\nI0930 16:12:07.475548 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0930 16:12:07.475622 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0930 16:12:07.475797 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0930 16:12:07.475804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://caff708b9ea16241fd0d989ee1a90e7a011ada229475d3993196566084913dc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://392ad0066e3794f60f9ce224c754743cc5360c7e3802dc05e258c128c9e69638\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.826364 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d650844-e53c-41d8-bcf1-fd11da30b7bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b24245811dc8414946d551978d02a35942824d4a2a9bb8a5d77a67c1caeaab55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428274e2745ac57e784fb4d1a88d140f38c68b568e6f80367bbb486495664a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c97c3e47b7b461af34ae7686ab62f02008de7d2b3dbbc795eb115141ae5f24e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f1a375db21aa81c00b5cc608d0c78af17002886bfa8a16ca16378d40ebe429\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.849610 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hldp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:13:02Z\\\",\\\"message\\\":\\\"2025-09-30T16:12:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d2e7a03b-5b56-4681-a4de-b8727b3d3fce\\\\n2025-09-30T16:12:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d2e7a03b-5b56-4681-a4de-b8727b3d3fce to /host/opt/cni/bin/\\\\n2025-09-30T16:12:17Z [verbose] multus-daemon started\\\\n2025-09-30T16:12:17Z [verbose] Readiness Indicator file check\\\\n2025-09-30T16:13:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6z6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hldp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.870543 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20a0a11f-4c9c-4412-96d8-57ad90ca6431\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95cd02ca7c1442a0694cf1f9e5d75139bc6c18c5ba6e89d934d41cd15ada4b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8be6d6dd37bcb003859caa96144b763571051b7132d13f0b90959040a08d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c6a2848c7716dd7cf0702246e1fc27f6d177b3337dc4a10f7dc6182671140b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f695b55bc0bdd15070b78945900514a2be4a33e2a5160c042ee072626e9b43ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.885248 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"080efbaf-9ef0-4538-aab9-624add222ac2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48af73cc5e378a96747a37927f934fb8f6591ff48a799d4cde3164a0a82053b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:11:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b7863ad82bf97f8113ebd8eb1a7bd09b2d15da5e9f1d93872690754cd282c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:11:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.906737 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://249f6757a05d6c6ececb6acd81056b678acd04d18b2c07c1cc1d3d56749cd9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.908561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.908601 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.908617 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.908640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.908661 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:08Z","lastTransitionTime":"2025-09-30T16:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.931318 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jskcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"414dbf05-2d31-4c78-aad6-6c8f890f87ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173bc04a314e91164d40ac1e06efe98cd1e1ad0d75e3fa940507e1e429ecc3b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cb7eb24e537f77174d8384bd8f70294c7a33bea55a7d38789dccb201fd95cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6b0c1e78795abf37fd5b8ee73b92eaa5d1ae8d51aa0af9545c0723d99ca52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79e52c43ca2b044aa7bb251712e8df841ea2eeee1382779175ade38f8139df3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a25735e18743b0f1e0c7a94223f865aa9cd6a2466aa38379183d49c487a977b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de7546877793e85d1733ce14ca8e1d6c5bd9d53f3d947e7049cc3a622a403ed9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa552efedd5ca04d61b3d6cd72e79351ec7c10fc36d306a466e6e622a23d7525\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cgzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jskcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.948909 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xtn5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7530dcc2-0e7d-41df-b43c-13a1a28cee98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8347a0ebd8f8ab85aac8118a3f3fa882680989c7b4db3c4d034335c2eef0473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2s8nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xtn5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.968302 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75a4b175-f2f2-4770-8463-7dba302f15c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b477704a19fe7b9ac07383f55b1ecf8d0a132de30226fc3dba7de35d6b2489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119c1358b84a71e0e7380122418d05c7e29243a98959e8ac397ad454054787fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8t7tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmzf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:08 crc kubenswrapper[4796]: I0930 16:13:08.988177 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1723e2684a8e1d5494cf0ec248b59f522f8f7da15c222aaa844be65e53491d43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:08Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.009323 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.017561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.017826 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.017851 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.017887 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.017912 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.028189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gpcn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f62bd4b-65ac-4b18-80ee-644b193c1912\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6129d74fffcb89d77328fa5aca1cd98ca669a235855bb11e89127f47482b8bc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z559s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gpcn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.061276 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7df6e69-ff3a-484a-b440-8afe12028b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T16:12:47Z\\\",\\\"message\\\":\\\"ector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 16:12:47.552049 6527 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.552270 6527 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 16:12:47.553693 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 16:12:47.553742 6527 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 16:12:47.553831 6527 factory.go:656] Stopping watch factory\\\\nI0930 16:12:47.553859 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 16:12:47.553825 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 16:12:47.572815 6527 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0930 16:12:47.572857 6527 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0930 16:12:47.572927 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0930 16:12:47.572957 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0930 16:12:47.573104 6527 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c6n4h_openshift-ovn-kubernetes(f7df6e69-ff3a-484a-b440-8afe12028b40)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmcms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c6n4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.081276 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g5mv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tl9z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.105064 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.120714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.120781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.120804 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.120833 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.120858 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.124845 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e66e5d6c0267c3a74978314ff97dc5094e6a3a88a324609d643a459c9ad96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47376252e06f3dfeb26eec5e43117b498439f0bc0e1a67140a911e989445f7c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.144451 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"670c655e-3953-4773-84ef-19c678d482f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T16:12:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375f748c57080345fa1a53334fe054802a1d8971a81fc637bcb723c6c1ba3945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T16:12:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vch6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T16:12:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8bbws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:09Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.223434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.223482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.223497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.223515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.223526 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.327268 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.327328 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.327346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.327399 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.327420 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.429641 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.429684 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.429694 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.429708 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.429719 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.533410 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.533475 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.533494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.533520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.533538 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.636691 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.636757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.636774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.636801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.636820 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.733224 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.733280 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.733276 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:09 crc kubenswrapper[4796]: E0930 16:13:09.733430 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:09 crc kubenswrapper[4796]: E0930 16:13:09.733547 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:09 crc kubenswrapper[4796]: E0930 16:13:09.733642 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.739682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.739736 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.739748 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.739770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.739783 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.843525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.843586 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.843819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.843845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.843866 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.946033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.946145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.946165 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.946188 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:09 crc kubenswrapper[4796]: I0930 16:13:09.946206 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:09Z","lastTransitionTime":"2025-09-30T16:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.049130 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.049182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.049197 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.049218 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.049233 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.152508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.152571 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.152583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.152601 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.152621 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.255463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.255496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.255507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.255524 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.255537 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.358614 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.358678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.358696 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.358720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.358736 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.461602 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.461674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.461701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.461726 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.461745 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.565512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.565628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.565667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.565717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.565757 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.669059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.669512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.669666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.669809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.669938 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.733271 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:10 crc kubenswrapper[4796]: E0930 16:13:10.733497 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.773131 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.773208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.773231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.773265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.773287 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.875903 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.876014 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.876033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.876059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.876076 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.979127 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.979199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.979217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.979242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:10 crc kubenswrapper[4796]: I0930 16:13:10.979260 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:10Z","lastTransitionTime":"2025-09-30T16:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.082587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.082644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.082661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.082687 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.082710 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.186718 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.186798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.186817 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.186842 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.186859 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.289244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.289284 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.289298 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.289314 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.289330 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.392223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.392263 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.392276 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.392293 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.392305 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.494832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.494888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.494905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.494927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.494945 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.552675 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.552830 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:14:15.552796179 +0000 UTC m=+147.566074736 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.553243 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.553366 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.553377 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.553643 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:14:15.553624923 +0000 UTC m=+147.566903450 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.553540 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.553858 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 16:14:15.55384799 +0000 UTC m=+147.567126587 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.597723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.597780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.597798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.597822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.597840 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.654236 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.654325 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654523 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654550 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654609 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654626 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654694 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 16:14:15.65467528 +0000 UTC m=+147.667953827 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654556 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654798 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.654876 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 16:14:15.654851475 +0000 UTC m=+147.668130032 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.700903 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.700967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.701013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.701042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.701059 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.732752 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.732878 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.733110 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.733178 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.733280 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:11 crc kubenswrapper[4796]: E0930 16:13:11.733481 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.803523 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.803557 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.803566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.803579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.803588 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.906932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.907045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.907068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.907121 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:11 crc kubenswrapper[4796]: I0930 16:13:11.907144 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:11Z","lastTransitionTime":"2025-09-30T16:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.010812 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.010868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.010885 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.010905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.010920 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.114192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.114316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.114334 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.114360 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.114378 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.217720 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.217777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.217789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.217808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.217819 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.320337 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.320411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.320434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.320464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.320489 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.424401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.424471 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.424489 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.424515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.424532 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.527507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.527574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.527597 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.527626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.527647 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.630918 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.631033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.631055 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.631080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.631096 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.732346 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:12 crc kubenswrapper[4796]: E0930 16:13:12.732580 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.734018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.734041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.734050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.734063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.734074 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.837692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.837755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.837776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.837807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.837828 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.940574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.940615 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.940630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.940649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:12 crc kubenswrapper[4796]: I0930 16:13:12.940664 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:12Z","lastTransitionTime":"2025-09-30T16:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.044881 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.044939 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.044957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.045013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.045035 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.148444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.148502 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.148518 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.148541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.148558 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.250515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.250559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.250573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.250589 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.250601 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.353585 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.353633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.353644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.353660 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.353672 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.456703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.456778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.456802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.456830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.456854 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.559418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.559485 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.559503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.559529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.559551 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.663737 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.663859 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.663885 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.663915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.663938 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.732830 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.732926 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.732847 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:13 crc kubenswrapper[4796]: E0930 16:13:13.733061 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:13 crc kubenswrapper[4796]: E0930 16:13:13.733170 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:13 crc kubenswrapper[4796]: E0930 16:13:13.733280 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.766682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.766721 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.766738 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.766760 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.766777 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.869257 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.869345 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.869373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.869399 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.869418 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.972844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.972915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.972939 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.972970 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:13 crc kubenswrapper[4796]: I0930 16:13:13.973036 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:13Z","lastTransitionTime":"2025-09-30T16:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.075300 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.075346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.075637 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.075668 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.075684 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.179019 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.179080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.179094 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.179139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.179154 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.282397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.282452 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.282469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.282490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.282509 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.385331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.385398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.385412 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.385437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.385498 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.489286 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.489370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.489388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.489414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.489434 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.592573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.592619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.592638 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.592657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.592671 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.695070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.695125 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.695143 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.695176 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.695193 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.732971 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:14 crc kubenswrapper[4796]: E0930 16:13:14.733214 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.734781 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.798149 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.798205 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.798223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.798247 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.798308 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.901302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.901368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.901385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.901412 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:14 crc kubenswrapper[4796]: I0930 16:13:14.901430 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:14Z","lastTransitionTime":"2025-09-30T16:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.004711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.004752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.004769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.004791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.004807 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.107230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.107662 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.107818 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.107939 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.108121 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.210825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.210871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.210883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.210899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.210912 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.313539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.313605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.313623 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.313648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.313665 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.417038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.417106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.417123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.417147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.417164 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.519837 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.519888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.519907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.519931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.519949 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.592618 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.592710 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.592728 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.592757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.592781 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.618179 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.631356 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.631384 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.631395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.631413 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.631426 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.656551 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.661826 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.661875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.661888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.661906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.661915 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.674218 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.678430 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.678453 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.678463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.678474 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.678481 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.690831 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.694906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.694941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.694953 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.694968 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.694994 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.711245 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T16:13:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6af019d7-e59f-45e8-81c8-64578033eb5e\\\",\\\"systemUUID\\\":\\\"880be81d-8176-4d82-bbcc-7517e59a48bd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T16:13:15Z is after 2025-08-24T17:21:41Z" Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.711380 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.713017 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.713066 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.713083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.713105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.713120 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.732581 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.732741 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.732778 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.732809 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.733117 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:15 crc kubenswrapper[4796]: E0930 16:13:15.733389 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.815806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.815881 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.815898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.815927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.815947 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.918616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.918680 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.918700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.918750 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:15 crc kubenswrapper[4796]: I0930 16:13:15.918768 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:15Z","lastTransitionTime":"2025-09-30T16:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.021365 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.021438 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.021458 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.021489 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.021513 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.124101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.124142 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.124154 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.124171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.124184 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.226472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.226515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.226526 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.226551 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.226563 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.245455 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/2.log" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.248298 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerStarted","Data":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.248922 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.275923 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.275898134 podStartE2EDuration="40.275898134s" podCreationTimestamp="2025-09-30 16:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.275713429 +0000 UTC m=+88.288991966" watchObservedRunningTime="2025-09-30 16:13:16.275898134 +0000 UTC m=+88.289176701" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.312392 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=22.312364646 podStartE2EDuration="22.312364646s" podCreationTimestamp="2025-09-30 16:12:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.288454213 +0000 UTC m=+88.301732770" watchObservedRunningTime="2025-09-30 16:13:16.312364646 +0000 UTC m=+88.325643213" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.329147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.329190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.329201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.329221 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.329233 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.338207 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-hldp6" podStartSLOduration=62.338190614 podStartE2EDuration="1m2.338190614s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.336826905 +0000 UTC m=+88.350105472" watchObservedRunningTime="2025-09-30 16:13:16.338190614 +0000 UTC m=+88.351469151" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.427716 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-gpcn7" podStartSLOduration=62.42769098 podStartE2EDuration="1m2.42769098s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.412245559 +0000 UTC m=+88.425524086" watchObservedRunningTime="2025-09-30 16:13:16.42769098 +0000 UTC m=+88.440969527" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.428130 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jskcj" podStartSLOduration=62.428124193 podStartE2EDuration="1m2.428124193s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.427510335 +0000 UTC m=+88.440788862" watchObservedRunningTime="2025-09-30 16:13:16.428124193 +0000 UTC m=+88.441402730" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.431806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.431869 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.431882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.431899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.431911 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.450896 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xtn5f" podStartSLOduration=62.450873973 podStartE2EDuration="1m2.450873973s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.438970273 +0000 UTC m=+88.452248820" watchObservedRunningTime="2025-09-30 16:13:16.450873973 +0000 UTC m=+88.464152500" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.465950 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmzf7" podStartSLOduration=61.465930563 podStartE2EDuration="1m1.465930563s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.451035417 +0000 UTC m=+88.464313944" watchObservedRunningTime="2025-09-30 16:13:16.465930563 +0000 UTC m=+88.479209100" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.521944 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podStartSLOduration=62.521925382 podStartE2EDuration="1m2.521925382s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.497401352 +0000 UTC m=+88.510679879" watchObservedRunningTime="2025-09-30 16:13:16.521925382 +0000 UTC m=+88.535203909" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.522148 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podStartSLOduration=62.522135718 podStartE2EDuration="1m2.522135718s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.521560112 +0000 UTC m=+88.534838639" watchObservedRunningTime="2025-09-30 16:13:16.522135718 +0000 UTC m=+88.535414245" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.533598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.533638 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.533646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.533658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.533675 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.571753 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.571730455 podStartE2EDuration="1m9.571730455s" podCreationTimestamp="2025-09-30 16:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.57154143 +0000 UTC m=+88.584819957" watchObservedRunningTime="2025-09-30 16:13:16.571730455 +0000 UTC m=+88.585008992" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.572262 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=67.57225687 podStartE2EDuration="1m7.57225687s" podCreationTimestamp="2025-09-30 16:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.55650927 +0000 UTC m=+88.569787797" watchObservedRunningTime="2025-09-30 16:13:16.57225687 +0000 UTC m=+88.585535397" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.598694 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.598675575 podStartE2EDuration="1m4.598675575s" podCreationTimestamp="2025-09-30 16:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:16.586124266 +0000 UTC m=+88.599402803" watchObservedRunningTime="2025-09-30 16:13:16.598675575 +0000 UTC m=+88.611954102" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.629465 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tl9z9"] Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.629602 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:16 crc kubenswrapper[4796]: E0930 16:13:16.629700 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.635648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.635685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.635695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.635710 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.635721 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.738266 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.738713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.738721 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.738736 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.738745 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.840894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.840931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.840941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.840957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.840966 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.943344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.943429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.943439 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.943456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:16 crc kubenswrapper[4796]: I0930 16:13:16.943467 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:16Z","lastTransitionTime":"2025-09-30T16:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.046740 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.046868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.046890 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.046916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.046940 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.149673 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.149723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.149738 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.149761 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.149775 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.252233 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.252304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.252327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.252355 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.252376 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.354566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.354614 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.354628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.354645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.354658 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.458001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.458041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.458052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.458068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.458080 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.562106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.562166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.562190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.562220 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.562241 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.665654 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.665709 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.665730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.665753 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.665771 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.733250 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.733405 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:17 crc kubenswrapper[4796]: E0930 16:13:17.733498 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.733267 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:17 crc kubenswrapper[4796]: E0930 16:13:17.733624 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 16:13:17 crc kubenswrapper[4796]: E0930 16:13:17.733790 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.768201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.768253 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.768265 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.768283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.768295 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.871434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.871481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.871492 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.871508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.871519 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.974640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.974690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.974700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.974717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:17 crc kubenswrapper[4796]: I0930 16:13:17.974728 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:17Z","lastTransitionTime":"2025-09-30T16:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.078889 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.079193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.079221 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.079258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.079328 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.182416 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.182460 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.182470 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.182488 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.182499 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.285781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.285837 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.285850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.285869 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.285882 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.388834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.388880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.388898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.388920 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.388937 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.491464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.491524 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.491540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.491558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.491570 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.594504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.594537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.594561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.594577 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.594585 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.696909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.696967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.696998 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.697024 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.697041 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.733098 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:18 crc kubenswrapper[4796]: E0930 16:13:18.734138 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tl9z9" podUID="d1683cfa-0a4a-4f40-bbbf-6675107ecbf9" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.800589 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.800656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.800669 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.800688 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.800698 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.903343 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.903384 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.903396 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.903409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:18 crc kubenswrapper[4796]: I0930 16:13:18.903419 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:18Z","lastTransitionTime":"2025-09-30T16:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.006826 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.006880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.006897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.006921 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.006941 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:19Z","lastTransitionTime":"2025-09-30T16:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.110120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.110186 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.110201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.110223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.110239 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:19Z","lastTransitionTime":"2025-09-30T16:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.213371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.213431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.213448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.213475 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.213494 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T16:13:19Z","lastTransitionTime":"2025-09-30T16:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.316089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.316498 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.316512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.316529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.316634 4796 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.362554 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5zv48"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.363698 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.365673 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.366722 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.366869 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.367016 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.368223 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.369543 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.369936 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.370328 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.370464 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.370765 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.372769 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.373378 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.373833 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.374488 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.375102 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.375427 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.375622 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.378510 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6svgm"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.378799 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.380592 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.380929 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.381154 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.383861 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.384953 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.385203 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.392819 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.393269 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.393787 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.395606 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjzqr"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.395894 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.395965 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mqjb"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.396478 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.397005 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.397183 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.398136 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.398190 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dxc8p"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.398502 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.398704 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.398893 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.398975 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.399068 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.399592 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.399714 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.399872 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.399973 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.400084 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.400484 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.403807 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.411327 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.415316 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.415466 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.416267 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.419481 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430237 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430289 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430310 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pbnlg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430531 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430603 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430678 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.430798 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.431009 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.431069 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.431108 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.431485 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.431589 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432022 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432064 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432118 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432194 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432363 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432449 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432543 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432627 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432666 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432766 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.432908 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.433277 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.434118 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-cch5s"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.434453 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.434710 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.434887 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.434962 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.435025 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.435105 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.435238 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.435358 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fl5fz"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.436993 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.437118 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.437266 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.437424 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.438017 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.438023 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.443842 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzwx5\" (UniqueName: \"kubernetes.io/projected/8adba875-c47a-49d0-a52e-5ddc97bfb928-kube-api-access-tzwx5\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.443927 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-serving-cert\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.443948 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plbmh\" (UniqueName: \"kubernetes.io/projected/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-kube-api-access-plbmh\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.443963 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/06101703-e6d8-42e5-8549-7340fbf1a597-node-pullsecrets\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.443991 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0baddf30-b72f-4659-b221-400c7b11ad0c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rk8v2\" (UID: \"0baddf30-b72f-4659-b221-400c7b11ad0c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444009 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444023 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-image-import-ca\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444045 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-etcd-client\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444060 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krvrl\" (UniqueName: \"kubernetes.io/projected/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-kube-api-access-krvrl\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444078 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8adba875-c47a-49d0-a52e-5ddc97bfb928-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444091 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5zf4\" (UniqueName: \"kubernetes.io/projected/06101703-e6d8-42e5-8549-7340fbf1a597-kube-api-access-z5zf4\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444107 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444147 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8adba875-c47a-49d0-a52e-5ddc97bfb928-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444162 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-serving-cert\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444178 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-config\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444191 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-serving-cert\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444207 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-config\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8adba875-c47a-49d0-a52e-5ddc97bfb928-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444237 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-service-ca-bundle\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444251 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-etcd-serving-ca\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444266 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-client-ca\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444288 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06101703-e6d8-42e5-8549-7340fbf1a597-audit-dir\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444304 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-config\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444317 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-469bd\" (UniqueName: \"kubernetes.io/projected/0baddf30-b72f-4659-b221-400c7b11ad0c-kube-api-access-469bd\") pod \"cluster-samples-operator-665b6dd947-rk8v2\" (UID: \"0baddf30-b72f-4659-b221-400c7b11ad0c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444332 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-audit\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444346 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-encryption-config\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.444652 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.445156 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ggvd5"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.445592 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.445775 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.446573 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.446745 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.446864 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h5scx"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.447560 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.448033 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.448568 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.448757 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.449346 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.451550 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.451562 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.462907 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.464200 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.464499 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.464861 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.465188 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.465648 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.465783 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.465941 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.466391 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.466788 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.467091 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.467807 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.468676 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.469258 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.469678 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.469908 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.470105 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.470234 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.470389 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.470570 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.470696 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.470821 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.471049 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.471225 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.471403 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.471441 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.471464 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.472555 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.483089 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.483186 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.483255 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.483617 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.485144 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.485250 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.485746 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-6r92c"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.487272 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.487750 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.488703 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.491601 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.491913 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.492297 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.492608 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.492758 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.492940 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.493149 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.493497 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.493557 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.493745 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.494006 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mvv7h"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.494119 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.494833 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.495406 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.495805 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.496023 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.496115 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.496419 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.496883 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.497152 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbg46"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.497853 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.499941 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.501309 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.504368 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s9qpw"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.504942 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.505254 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.505831 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.506856 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.507349 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.507537 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.507667 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h6hqm"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.508222 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.511199 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.512015 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.512373 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.513114 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.513632 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5zv48"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.517235 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.527645 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.530470 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.530753 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cch5s"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.531388 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h5scx"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.533160 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pbnlg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.534098 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-j6mpg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.536035 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dxc8p"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.536217 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.537473 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-sl6k6"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.538175 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.538685 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.565809 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566064 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566725 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8adba875-c47a-49d0-a52e-5ddc97bfb928-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566762 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-service-ca\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566788 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5zf4\" (UniqueName: \"kubernetes.io/projected/06101703-e6d8-42e5-8549-7340fbf1a597-kube-api-access-z5zf4\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566814 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73414d05-f9c7-49fc-9770-d12886f12382-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566837 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.566860 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/290a8e80-8ade-4087-84f3-8223e380c0ef-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567107 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4c6f0a6e-14ff-45a7-8312-09fd13121728-machine-approver-tls\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567161 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567288 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-oauth-serving-cert\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567328 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/290a8e80-8ade-4087-84f3-8223e380c0ef-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567369 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-config\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567391 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f36e801-bfb5-425f-876d-ed8191d5e66e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567452 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52l66\" (UniqueName: \"kubernetes.io/projected/2434945c-60ef-40c1-aa33-2cded03cc0e0-kube-api-access-52l66\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567478 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9f9b\" (UniqueName: \"kubernetes.io/projected/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-kube-api-access-t9f9b\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567501 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567770 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.567925 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568158 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568204 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73dd5dd2-4416-4804-9377-e66521991a5e-serving-cert\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568224 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q47m\" (UniqueName: \"kubernetes.io/projected/1b2cdffa-2549-4fb3-aa14-6b2da0738146-kube-api-access-6q47m\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568247 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-ca\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568272 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-oauth-config\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568298 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8adba875-c47a-49d0-a52e-5ddc97bfb928-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568361 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldd7x\" (UniqueName: \"kubernetes.io/projected/025a66cf-d82d-4590-b9cd-402bc6bc3668-kube-api-access-ldd7x\") pod \"downloads-7954f5f757-dxc8p\" (UID: \"025a66cf-d82d-4590-b9cd-402bc6bc3668\") " pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568381 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c6f0a6e-14ff-45a7-8312-09fd13121728-auth-proxy-config\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568400 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch5zm\" (UniqueName: \"kubernetes.io/projected/4c6f0a6e-14ff-45a7-8312-09fd13121728-kube-api-access-ch5zm\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568419 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568446 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-serving-cert\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.568465 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kzpb\" (UniqueName: \"kubernetes.io/projected/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-kube-api-access-7kzpb\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570008 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570060 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570119 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-config\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570141 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-serving-cert\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570190 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570214 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-config\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570258 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd5dd2-4416-4804-9377-e66521991a5e-config\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570279 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-trusted-ca-bundle\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570529 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-dir\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.570966 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571047 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8adba875-c47a-49d0-a52e-5ddc97bfb928-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-policies\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571134 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-stats-auth\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571395 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8adba875-c47a-49d0-a52e-5ddc97bfb928-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571414 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-config\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571489 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-service-ca-bundle\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571526 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-etcd-serving-ca\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571554 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b2cdffa-2549-4fb3-aa14-6b2da0738146-serving-cert\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571574 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/290a8e80-8ade-4087-84f3-8223e380c0ef-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-service-ca\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571699 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6f0a6e-14ff-45a7-8312-09fd13121728-config\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571756 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-client\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571785 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f36e801-bfb5-425f-876d-ed8191d5e66e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571806 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-srv-cert\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571881 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwsrd\" (UniqueName: \"kubernetes.io/projected/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-kube-api-access-nwsrd\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571931 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-client-ca\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.571961 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06101703-e6d8-42e5-8549-7340fbf1a597-audit-dir\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572031 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572039 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06101703-e6d8-42e5-8549-7340fbf1a597-audit-dir\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572081 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73414d05-f9c7-49fc-9770-d12886f12382-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572128 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-config\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572216 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-469bd\" (UniqueName: \"kubernetes.io/projected/0baddf30-b72f-4659-b221-400c7b11ad0c-kube-api-access-469bd\") pod \"cluster-samples-operator-665b6dd947-rk8v2\" (UID: \"0baddf30-b72f-4659-b221-400c7b11ad0c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-serving-cert\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572282 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f36e801-bfb5-425f-876d-ed8191d5e66e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572314 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-audit\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572342 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-encryption-config\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572480 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzwx5\" (UniqueName: \"kubernetes.io/projected/8adba875-c47a-49d0-a52e-5ddc97bfb928-kube-api-access-tzwx5\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572572 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73dd5dd2-4416-4804-9377-e66521991a5e-trusted-ca\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572607 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-default-certificate\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572611 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-service-ca-bundle\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572631 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-serving-cert\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572652 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/06101703-e6d8-42e5-8549-7340fbf1a597-node-pullsecrets\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572676 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plbmh\" (UniqueName: \"kubernetes.io/projected/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-kube-api-access-plbmh\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572718 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0baddf30-b72f-4659-b221-400c7b11ad0c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rk8v2\" (UID: \"0baddf30-b72f-4659-b221-400c7b11ad0c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572727 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-client-ca\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572737 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7gbh\" (UniqueName: \"kubernetes.io/projected/73dd5dd2-4416-4804-9377-e66521991a5e-kube-api-access-b7gbh\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572729 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-config\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572759 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-metrics-certs\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572806 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkbt6\" (UniqueName: \"kubernetes.io/projected/73414d05-f9c7-49fc-9770-d12886f12382-kube-api-access-vkbt6\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572826 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/06101703-e6d8-42e5-8549-7340fbf1a597-node-pullsecrets\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.572926 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573103 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-config\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573156 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-image-import-ca\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573252 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krvrl\" (UniqueName: \"kubernetes.io/projected/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-kube-api-access-krvrl\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573287 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/290a8e80-8ade-4087-84f3-8223e380c0ef-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573355 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-audit\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573377 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/290a8e80-8ade-4087-84f3-8223e380c0ef-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573621 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-etcd-client\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573676 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-config\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573705 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573823 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-service-ca-bundle\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.573941 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-image-import-ca\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.574012 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.574440 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/06101703-e6d8-42e5-8549-7340fbf1a597-etcd-serving-ca\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.574688 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.577704 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0baddf30-b72f-4659-b221-400c7b11ad0c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rk8v2\" (UID: \"0baddf30-b72f-4659-b221-400c7b11ad0c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.580399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-serving-cert\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.580477 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-serving-cert\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.580511 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.580415 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-encryption-config\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.580828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8adba875-c47a-49d0-a52e-5ddc97bfb928-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.582093 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.582815 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mqjb"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.584434 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-serving-cert\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.584568 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fl5fz"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.585783 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.586573 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6svgm"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.587592 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.588552 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.589507 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.590424 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.591395 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.592308 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mvv7h"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.593474 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.594443 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/06101703-e6d8-42e5-8549-7340fbf1a597-etcd-client\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.594703 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbg46"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.595772 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.597270 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.599172 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ggvd5"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.601291 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.602648 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-mlhmv"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.603437 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.603788 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.605533 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.606266 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fw6t6"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.608076 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjzqr"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.608200 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.608453 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.610137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s9qpw"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.611276 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.612324 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mlhmv"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.613293 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h6hqm"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.614255 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.615232 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.616281 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-j6mpg"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.617274 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.618233 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fw6t6"] Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.625849 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.646572 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.666579 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.674933 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-default-certificate\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.674968 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675005 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7gbh\" (UniqueName: \"kubernetes.io/projected/73dd5dd2-4416-4804-9377-e66521991a5e-kube-api-access-b7gbh\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675024 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-metrics-certs\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675040 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkbt6\" (UniqueName: \"kubernetes.io/projected/73414d05-f9c7-49fc-9770-d12886f12382-kube-api-access-vkbt6\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675062 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/290a8e80-8ade-4087-84f3-8223e380c0ef-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675077 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/290a8e80-8ade-4087-84f3-8223e380c0ef-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675100 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-service-ca-bundle\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675118 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-config\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675134 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675156 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-service-ca\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675178 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73414d05-f9c7-49fc-9770-d12886f12382-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675194 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/290a8e80-8ade-4087-84f3-8223e380c0ef-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675209 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4c6f0a6e-14ff-45a7-8312-09fd13121728-machine-approver-tls\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675225 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-oauth-serving-cert\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675240 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675256 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675284 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/290a8e80-8ade-4087-84f3-8223e380c0ef-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675301 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-config\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675318 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f36e801-bfb5-425f-876d-ed8191d5e66e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675336 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52l66\" (UniqueName: \"kubernetes.io/projected/2434945c-60ef-40c1-aa33-2cded03cc0e0-kube-api-access-52l66\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675353 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9f9b\" (UniqueName: \"kubernetes.io/projected/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-kube-api-access-t9f9b\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675370 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675383 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/290a8e80-8ade-4087-84f3-8223e380c0ef-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675396 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675410 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73dd5dd2-4416-4804-9377-e66521991a5e-serving-cert\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675425 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q47m\" (UniqueName: \"kubernetes.io/projected/1b2cdffa-2549-4fb3-aa14-6b2da0738146-kube-api-access-6q47m\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675440 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-ca\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675457 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldd7x\" (UniqueName: \"kubernetes.io/projected/025a66cf-d82d-4590-b9cd-402bc6bc3668-kube-api-access-ldd7x\") pod \"downloads-7954f5f757-dxc8p\" (UID: \"025a66cf-d82d-4590-b9cd-402bc6bc3668\") " pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675471 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-oauth-config\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675489 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kzpb\" (UniqueName: \"kubernetes.io/projected/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-kube-api-access-7kzpb\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675502 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c6f0a6e-14ff-45a7-8312-09fd13121728-auth-proxy-config\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675519 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch5zm\" (UniqueName: \"kubernetes.io/projected/4c6f0a6e-14ff-45a7-8312-09fd13121728-kube-api-access-ch5zm\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675535 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675556 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675571 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675589 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675604 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-trusted-ca-bundle\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675620 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-dir\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675636 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd5dd2-4416-4804-9377-e66521991a5e-config\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675652 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-policies\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675667 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675683 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-stats-auth\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675699 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b2cdffa-2549-4fb3-aa14-6b2da0738146-serving-cert\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675715 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-service-ca\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675731 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6f0a6e-14ff-45a7-8312-09fd13121728-config\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675750 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/290a8e80-8ade-4087-84f3-8223e380c0ef-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675772 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f36e801-bfb5-425f-876d-ed8191d5e66e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675792 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-client\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-srv-cert\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675830 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwsrd\" (UniqueName: \"kubernetes.io/projected/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-kube-api-access-nwsrd\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675861 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675890 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73414d05-f9c7-49fc-9770-d12886f12382-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675913 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-serving-cert\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675935 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f36e801-bfb5-425f-876d-ed8191d5e66e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.675972 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73dd5dd2-4416-4804-9377-e66521991a5e-trusted-ca\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.676574 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.676881 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73414d05-f9c7-49fc-9770-d12886f12382-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.677194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-config\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.677331 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-service-ca\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.677674 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-service-ca\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.678283 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.678408 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-dir\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.678906 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/290a8e80-8ade-4087-84f3-8223e380c0ef-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.679229 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6f0a6e-14ff-45a7-8312-09fd13121728-config\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.679284 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/290a8e80-8ade-4087-84f3-8223e380c0ef-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.679309 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-trusted-ca-bundle\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.679753 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c6f0a6e-14ff-45a7-8312-09fd13121728-auth-proxy-config\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.679831 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-ca\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.680099 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.680424 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.680525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f36e801-bfb5-425f-876d-ed8191d5e66e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.680775 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.680922 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.681560 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-config\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.681584 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/290a8e80-8ade-4087-84f3-8223e380c0ef-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682259 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-policies\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682378 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4c6f0a6e-14ff-45a7-8312-09fd13121728-machine-approver-tls\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682698 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b2cdffa-2549-4fb3-aa14-6b2da0738146-etcd-client\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682814 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-oauth-serving-cert\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682859 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f36e801-bfb5-425f-876d-ed8191d5e66e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682866 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.682902 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-oauth-config\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.683408 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-serving-cert\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.683992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.684559 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73414d05-f9c7-49fc-9770-d12886f12382-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.684570 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.685176 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b2cdffa-2549-4fb3-aa14-6b2da0738146-serving-cert\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.686033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.686167 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.687260 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.715317 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.727226 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.732690 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.732916 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.732916 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.746805 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.766687 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.806206 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.826270 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.828438 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd5dd2-4416-4804-9377-e66521991a5e-config\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.845330 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.867398 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.875486 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73dd5dd2-4416-4804-9377-e66521991a5e-serving-cert\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.896516 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.905891 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.907047 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73dd5dd2-4416-4804-9377-e66521991a5e-trusted-ca\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.926939 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.945924 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.953360 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-srv-cert\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.966781 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.971597 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:19 crc kubenswrapper[4796]: I0930 16:13:19.987390 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.006773 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.026532 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.045945 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.051609 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-stats-auth\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.067538 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.079285 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-metrics-certs\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.087585 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.107545 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.119287 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-default-certificate\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.128138 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.147350 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.157669 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-service-ca-bundle\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.168268 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.188222 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.206760 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.227782 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.248142 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.267405 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.286529 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.306118 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.346789 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.367893 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.386895 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.407711 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.427328 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.447075 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.467416 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.487171 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.505461 4796 request.go:700] Waited for 1.007700677s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dservice-ca-operator-dockercfg-rg9jl&limit=500&resourceVersion=0 Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.508774 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.527904 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.548220 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.567301 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.586620 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.606465 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.625967 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.646736 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.666790 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.686866 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.707747 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.726605 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.733295 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.747392 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.781371 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.786278 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.806622 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.827081 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.846946 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.868274 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.887256 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.906660 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.926732 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.946114 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.967049 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 30 16:13:20 crc kubenswrapper[4796]: I0930 16:13:20.987746 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.007461 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.025753 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.046872 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.067212 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.086137 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.106660 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.126730 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.147623 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.166893 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.186935 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.206443 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.227211 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.247293 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.267146 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.286815 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.306509 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.327431 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.346625 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.366251 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.387303 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.423043 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8adba875-c47a-49d0-a52e-5ddc97bfb928-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.443228 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5zf4\" (UniqueName: \"kubernetes.io/projected/06101703-e6d8-42e5-8549-7340fbf1a597-kube-api-access-z5zf4\") pod \"apiserver-76f77b778f-5zv48\" (UID: \"06101703-e6d8-42e5-8549-7340fbf1a597\") " pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.464014 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzwx5\" (UniqueName: \"kubernetes.io/projected/8adba875-c47a-49d0-a52e-5ddc97bfb928-kube-api-access-tzwx5\") pod \"cluster-image-registry-operator-dc59b4c8b-v222h\" (UID: \"8adba875-c47a-49d0-a52e-5ddc97bfb928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.486447 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.499442 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-469bd\" (UniqueName: \"kubernetes.io/projected/0baddf30-b72f-4659-b221-400c7b11ad0c-kube-api-access-469bd\") pod \"cluster-samples-operator-665b6dd947-rk8v2\" (UID: \"0baddf30-b72f-4659-b221-400c7b11ad0c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.505124 4796 request.go:700] Waited for 1.931513057s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.516579 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plbmh\" (UniqueName: \"kubernetes.io/projected/6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f-kube-api-access-plbmh\") pod \"authentication-operator-69f744f599-6svgm\" (UID: \"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.528709 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.540735 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krvrl\" (UniqueName: \"kubernetes.io/projected/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-kube-api-access-krvrl\") pod \"route-controller-manager-6576b87f9c-6dvjn\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.544353 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.547137 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.565197 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.567907 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.596493 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.606468 4796 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.627837 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.668164 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7gbh\" (UniqueName: \"kubernetes.io/projected/73dd5dd2-4416-4804-9377-e66521991a5e-kube-api-access-b7gbh\") pod \"console-operator-58897d9998-h5scx\" (UID: \"73dd5dd2-4416-4804-9377-e66521991a5e\") " pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.686473 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkbt6\" (UniqueName: \"kubernetes.io/projected/73414d05-f9c7-49fc-9770-d12886f12382-kube-api-access-vkbt6\") pod \"openshift-apiserver-operator-796bbdcf4f-8cfhw\" (UID: \"73414d05-f9c7-49fc-9770-d12886f12382\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.702938 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/290a8e80-8ade-4087-84f3-8223e380c0ef-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sj2sq\" (UID: \"290a8e80-8ade-4087-84f3-8223e380c0ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.719421 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwsrd\" (UniqueName: \"kubernetes.io/projected/839c7aeb-ed29-4196-9fc5-0a8eecf7e585-kube-api-access-nwsrd\") pod \"olm-operator-6b444d44fb-gp486\" (UID: \"839c7aeb-ed29-4196-9fc5-0a8eecf7e585\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.740752 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kzpb\" (UniqueName: \"kubernetes.io/projected/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-kube-api-access-7kzpb\") pod \"console-f9d7485db-cch5s\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.758465 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.763289 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch5zm\" (UniqueName: \"kubernetes.io/projected/4c6f0a6e-14ff-45a7-8312-09fd13121728-kube-api-access-ch5zm\") pod \"machine-approver-56656f9798-w2nkl\" (UID: \"4c6f0a6e-14ff-45a7-8312-09fd13121728\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.764256 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h"] Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.766292 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.780218 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6svgm"] Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.782908 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q47m\" (UniqueName: \"kubernetes.io/projected/1b2cdffa-2549-4fb3-aa14-6b2da0738146-kube-api-access-6q47m\") pod \"etcd-operator-b45778765-ggvd5\" (UID: \"1b2cdffa-2549-4fb3-aa14-6b2da0738146\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:21 crc kubenswrapper[4796]: W0930 16:13:21.786303 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b8528d4_f5f0_4d0a_b7a6_2d283ef9d62f.slice/crio-34e794d2503c3dd87e1c374d8b13b06f8b77f61f2b4bcdcb1102137c6ccd1ed0 WatchSource:0}: Error finding container 34e794d2503c3dd87e1c374d8b13b06f8b77f61f2b4bcdcb1102137c6ccd1ed0: Status 404 returned error can't find the container with id 34e794d2503c3dd87e1c374d8b13b06f8b77f61f2b4bcdcb1102137c6ccd1ed0 Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.797523 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.800529 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldd7x\" (UniqueName: \"kubernetes.io/projected/025a66cf-d82d-4590-b9cd-402bc6bc3668-kube-api-access-ldd7x\") pod \"downloads-7954f5f757-dxc8p\" (UID: \"025a66cf-d82d-4590-b9cd-402bc6bc3668\") " pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.816123 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.825389 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52l66\" (UniqueName: \"kubernetes.io/projected/2434945c-60ef-40c1-aa33-2cded03cc0e0-kube-api-access-52l66\") pod \"oauth-openshift-558db77b4-9mqjb\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.843457 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f36e801-bfb5-425f-876d-ed8191d5e66e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l8rrg\" (UID: \"9f36e801-bfb5-425f-876d-ed8191d5e66e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.863086 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9f9b\" (UniqueName: \"kubernetes.io/projected/fb586c4b-79c4-4747-b9d4-8ff7e21e03f2-kube-api-access-t9f9b\") pod \"router-default-5444994796-6r92c\" (UID: \"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2\") " pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.867872 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.886853 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.888803 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.905619 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.922552 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.932223 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.933029 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.944972 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5zv48"] Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.956354 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h5scx"] Sep 30 16:13:21 crc kubenswrapper[4796]: W0930 16:13:21.959845 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06101703_e6d8_42e5_8549_7340fbf1a597.slice/crio-0d0a59ade7f0b317e112014e8c20f06162da7e4add2b1c1456371dc332f75a6c WatchSource:0}: Error finding container 0d0a59ade7f0b317e112014e8c20f06162da7e4add2b1c1456371dc332f75a6c: Status 404 returned error can't find the container with id 0d0a59ade7f0b317e112014e8c20f06162da7e4add2b1c1456371dc332f75a6c Sep 30 16:13:21 crc kubenswrapper[4796]: W0930 16:13:21.962651 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73dd5dd2_4416_4804_9377_e66521991a5e.slice/crio-2604e28d731428acfc5465562e5d6a058c375e33ca39e88e018651579db10b5e WatchSource:0}: Error finding container 2604e28d731428acfc5465562e5d6a058c375e33ca39e88e018651579db10b5e: Status 404 returned error can't find the container with id 2604e28d731428acfc5465562e5d6a058c375e33ca39e88e018651579db10b5e Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.986570 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.987864 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486"] Sep 30 16:13:21 crc kubenswrapper[4796]: I0930 16:13:21.999291 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.006671 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.008480 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-bound-sa-token\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010643 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml24h\" (UniqueName: \"kubernetes.io/projected/ecf5a09b-2113-4202-83c1-ad8c77520791-kube-api-access-ml24h\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010680 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-config\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010696 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5271ef96-e4c6-49a4-b75b-b7ddf035402a-serving-cert\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010714 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47d764ec-58e3-4604-a3b6-259dc3c3fa36-audit-dir\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010735 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49060193-d216-473c-8b67-6eb0d8970dab-trusted-ca\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010758 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010777 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-trusted-ca\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010791 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8aba7c6e-3d24-4217-9687-93a535675dbc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010805 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-etcd-client\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010818 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-serving-cert\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010831 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5sjs\" (UniqueName: \"kubernetes.io/projected/47d764ec-58e3-4604-a3b6-259dc3c3fa36-kube-api-access-j5sjs\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010862 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010879 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010895 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ecf5a09b-2113-4202-83c1-ad8c77520791-images\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czkqh\" (UniqueName: \"kubernetes.io/projected/49060193-d216-473c-8b67-6eb0d8970dab-kube-api-access-czkqh\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010935 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqf9p\" (UniqueName: \"kubernetes.io/projected/f2cf432c-1f44-4dd8-9f0b-c536db37b36b-kube-api-access-bqf9p\") pod \"migrator-59844c95c7-22cb7\" (UID: \"f2cf432c-1f44-4dd8-9f0b-c536db37b36b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010948 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/49060193-d216-473c-8b67-6eb0d8970dab-bound-sa-token\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.010963 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd4j8\" (UniqueName: \"kubernetes.io/projected/5271ef96-e4c6-49a4-b75b-b7ddf035402a-kube-api-access-jd4j8\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011007 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzt5s\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-kube-api-access-xzt5s\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011027 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7adf9170-1515-4c07-b471-bcc775bf4327-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011042 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8aba7c6e-3d24-4217-9687-93a535675dbc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011059 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-encryption-config\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf5a09b-2113-4202-83c1-ad8c77520791-config\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011095 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv7nz\" (UniqueName: \"kubernetes.io/projected/7adf9170-1515-4c07-b471-bcc775bf4327-kube-api-access-cv7nz\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-certificates\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011123 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-audit-policies\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011138 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-client-ca\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011153 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ecf5a09b-2113-4202-83c1-ad8c77520791-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011181 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011212 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-tls\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011228 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7adf9170-1515-4c07-b471-bcc775bf4327-serving-cert\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.011243 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49060193-d216-473c-8b67-6eb0d8970dab-metrics-tls\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.011596 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:22.511583665 +0000 UTC m=+94.524862192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.013668 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.030242 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.038170 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" Sep 30 16:13:22 crc kubenswrapper[4796]: W0930 16:13:22.058785 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c6f0a6e_14ff_45a7_8312_09fd13121728.slice/crio-2cced5da2bcdc9c69901de78b02df456841aab910c08efd4c6389bc8b89a394c WatchSource:0}: Error finding container 2cced5da2bcdc9c69901de78b02df456841aab910c08efd4c6389bc8b89a394c: Status 404 returned error can't find the container with id 2cced5da2bcdc9c69901de78b02df456841aab910c08efd4c6389bc8b89a394c Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.074726 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112670 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112842 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqf9p\" (UniqueName: \"kubernetes.io/projected/f2cf432c-1f44-4dd8-9f0b-c536db37b36b-kube-api-access-bqf9p\") pod \"migrator-59844c95c7-22cb7\" (UID: \"f2cf432c-1f44-4dd8-9f0b-c536db37b36b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/49060193-d216-473c-8b67-6eb0d8970dab-bound-sa-token\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112886 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd4j8\" (UniqueName: \"kubernetes.io/projected/5271ef96-e4c6-49a4-b75b-b7ddf035402a-kube-api-access-jd4j8\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112905 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/73c91ec0-f08c-4a00-8b90-a8443d548128-certs\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112939 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5j6m\" (UniqueName: \"kubernetes.io/projected/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-kube-api-access-l5j6m\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112956 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxqp9\" (UniqueName: \"kubernetes.io/projected/c239d916-4d86-47a9-b2c5-cac73bcea6bf-kube-api-access-qxqp9\") pod \"package-server-manager-789f6589d5-6xlsv\" (UID: \"c239d916-4d86-47a9-b2c5-cac73bcea6bf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.112996 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-csi-data-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113029 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-registration-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113045 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113077 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8aba7c6e-3d24-4217-9687-93a535675dbc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113091 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5496d724-4b33-4818-a317-d1ace3c64aeb-proxy-tls\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113108 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-signing-key\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113132 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-certificates\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113148 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a540a9ba-ecd7-4490-805b-45ff88d97a63-apiservice-cert\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113165 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55sc4\" (UniqueName: \"kubernetes.io/projected/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-kube-api-access-55sc4\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113191 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/309a1b03-85dc-466d-afdb-8c946fd030e1-serving-cert\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113205 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c5js\" (UniqueName: \"kubernetes.io/projected/309a1b03-85dc-466d-afdb-8c946fd030e1-kube-api-access-8c5js\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8czsm\" (UniqueName: \"kubernetes.io/projected/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-kube-api-access-8czsm\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.113261 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:22.61324671 +0000 UTC m=+94.626525237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113281 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/577762e0-d3ab-4927-8561-fdd79b43f5b1-profile-collector-cert\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113309 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113672 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/edd79356-997e-4b56-b884-921db4a45d38-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h6hqm\" (UID: \"edd79356-997e-4b56-b884-921db4a45d38\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113740 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/daf84026-ecfa-446d-91dc-e8146d350877-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.113774 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66tb\" (UniqueName: \"kubernetes.io/projected/577762e0-d3ab-4927-8561-fdd79b43f5b1-kube-api-access-f66tb\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.114203 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8aba7c6e-3d24-4217-9687-93a535675dbc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115026 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49060193-d216-473c-8b67-6eb0d8970dab-metrics-tls\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115057 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-bound-sa-token\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115147 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-signing-cabundle\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115165 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d34ff06-5475-4d4f-9e69-c0734d22554b-config\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115201 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9d4f\" (UniqueName: \"kubernetes.io/projected/15579146-527c-4348-9e39-6a9d0ba6533e-kube-api-access-t9d4f\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115220 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-images\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.115571 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.117719 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-certificates\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.117915 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47d764ec-58e3-4604-a3b6-259dc3c3fa36-audit-dir\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.117945 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65s4q\" (UniqueName: \"kubernetes.io/projected/a540a9ba-ecd7-4490-805b-45ff88d97a63-kube-api-access-65s4q\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.117994 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.118184 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47d764ec-58e3-4604-a3b6-259dc3c3fa36-audit-dir\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.118423 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.118874 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:22.61885419 +0000 UTC m=+94.632132717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119252 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-trusted-ca\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119281 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8aba7c6e-3d24-4217-9687-93a535675dbc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119339 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j94d5\" (UniqueName: \"kubernetes.io/projected/5496d724-4b33-4818-a317-d1ace3c64aeb-kube-api-access-j94d5\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119367 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-etcd-client\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119409 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f63ad2d2-d528-48e8-99f8-891d91f72ded-metrics-tls\") pod \"dns-operator-744455d44c-mvv7h\" (UID: \"f63ad2d2-d528-48e8-99f8-891d91f72ded\") " pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119477 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119506 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119529 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ecf5a09b-2113-4202-83c1-ad8c77520791-images\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.119644 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/be7f735b-ccce-4060-915d-58aea627d7f5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cwrnl\" (UID: \"be7f735b-ccce-4060-915d-58aea627d7f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121634 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-trusted-ca\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121748 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/73c91ec0-f08c-4a00-8b90-a8443d548128-node-bootstrap-token\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121778 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/309a1b03-85dc-466d-afdb-8c946fd030e1-config\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121814 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzt5s\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-kube-api-access-xzt5s\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121832 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhwjf\" (UniqueName: \"kubernetes.io/projected/73c91ec0-f08c-4a00-8b90-a8443d548128-kube-api-access-mhwjf\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121850 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g9v5\" (UniqueName: \"kubernetes.io/projected/f63ad2d2-d528-48e8-99f8-891d91f72ded-kube-api-access-9g9v5\") pod \"dns-operator-744455d44c-mvv7h\" (UID: \"f63ad2d2-d528-48e8-99f8-891d91f72ded\") " pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7adf9170-1515-4c07-b471-bcc775bf4327-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-encryption-config\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.121993 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf5a09b-2113-4202-83c1-ad8c77520791-config\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122011 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wq4z\" (UniqueName: \"kubernetes.io/projected/a355257c-502e-495a-97f0-bc65d5f1cfd5-kube-api-access-4wq4z\") pod \"ingress-canary-j6mpg\" (UID: \"a355257c-502e-495a-97f0-bc65d5f1cfd5\") " pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122044 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv7nz\" (UniqueName: \"kubernetes.io/projected/7adf9170-1515-4c07-b471-bcc775bf4327-kube-api-access-cv7nz\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122063 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-config-volume\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122250 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ecf5a09b-2113-4202-83c1-ad8c77520791-images\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122569 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7adf9170-1515-4c07-b471-bcc775bf4327-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122885 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122925 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a355257c-502e-495a-97f0-bc65d5f1cfd5-cert\") pod \"ingress-canary-j6mpg\" (UID: \"a355257c-502e-495a-97f0-bc65d5f1cfd5\") " pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122945 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-audit-policies\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.122996 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-client-ca\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.125588 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-etcd-client\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.126625 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8aba7c6e-3d24-4217-9687-93a535675dbc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.128823 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-audit-policies\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.129224 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-encryption-config\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.130965 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-client-ca\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.123061 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcrpg\" (UniqueName: \"kubernetes.io/projected/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-kube-api-access-pcrpg\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132521 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2p6v\" (UniqueName: \"kubernetes.io/projected/daf84026-ecfa-446d-91dc-e8146d350877-kube-api-access-n2p6v\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132544 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5jb5\" (UniqueName: \"kubernetes.io/projected/be7f735b-ccce-4060-915d-58aea627d7f5-kube-api-access-v5jb5\") pod \"control-plane-machine-set-operator-78cbb6b69f-cwrnl\" (UID: \"be7f735b-ccce-4060-915d-58aea627d7f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132560 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ecf5a09b-2113-4202-83c1-ad8c77520791-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132621 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-secret-volume\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132637 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5496d724-4b33-4818-a317-d1ace3c64aeb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132643 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf5a09b-2113-4202-83c1-ad8c77520791-config\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132654 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c239d916-4d86-47a9-b2c5-cac73bcea6bf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6xlsv\" (UID: \"c239d916-4d86-47a9-b2c5-cac73bcea6bf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.132948 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49060193-d216-473c-8b67-6eb0d8970dab-metrics-tls\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.133580 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47d764ec-58e3-4604-a3b6-259dc3c3fa36-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.135463 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-socket-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.135696 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.135871 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-tls\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.135970 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7adf9170-1515-4c07-b471-bcc775bf4327-serving-cert\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136054 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a540a9ba-ecd7-4490-805b-45ff88d97a63-tmpfs\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136084 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tplx\" (UniqueName: \"kubernetes.io/projected/edd79356-997e-4b56-b884-921db4a45d38-kube-api-access-5tplx\") pod \"multus-admission-controller-857f4d67dd-h6hqm\" (UID: \"edd79356-997e-4b56-b884-921db4a45d38\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d34ff06-5475-4d4f-9e69-c0734d22554b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136130 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-plugins-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136154 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmjjh\" (UniqueName: \"kubernetes.io/projected/889e3661-2ba6-45ed-a371-3d9b6300db64-kube-api-access-tmjjh\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136198 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml24h\" (UniqueName: \"kubernetes.io/projected/ecf5a09b-2113-4202-83c1-ad8c77520791-kube-api-access-ml24h\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136225 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15579146-527c-4348-9e39-6a9d0ba6533e-config-volume\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136292 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-config\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136317 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136359 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5271ef96-e4c6-49a4-b75b-b7ddf035402a-serving-cert\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136421 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-mountpoint-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.136462 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/577762e0-d3ab-4927-8561-fdd79b43f5b1-srv-cert\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138114 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49060193-d216-473c-8b67-6eb0d8970dab-trusted-ca\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138164 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a540a9ba-ecd7-4490-805b-45ff88d97a63-webhook-cert\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138182 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-proxy-tls\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138222 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-serving-cert\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138241 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5sjs\" (UniqueName: \"kubernetes.io/projected/47d764ec-58e3-4604-a3b6-259dc3c3fa36-kube-api-access-j5sjs\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138279 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf84026-ecfa-446d-91dc-e8146d350877-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138296 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15579146-527c-4348-9e39-6a9d0ba6533e-metrics-tls\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138311 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-config\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138352 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czkqh\" (UniqueName: \"kubernetes.io/projected/49060193-d216-473c-8b67-6eb0d8970dab-kube-api-access-czkqh\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138371 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p4s9\" (UniqueName: \"kubernetes.io/projected/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-kube-api-access-4p4s9\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.138388 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d34ff06-5475-4d4f-9e69-c0734d22554b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.139016 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.139879 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-config\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.142594 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49060193-d216-473c-8b67-6eb0d8970dab-trusted-ca\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.143107 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ecf5a09b-2113-4202-83c1-ad8c77520791-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.143335 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7adf9170-1515-4c07-b471-bcc775bf4327-serving-cert\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.144179 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-tls\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.144588 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d764ec-58e3-4604-a3b6-259dc3c3fa36-serving-cert\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.170363 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5271ef96-e4c6-49a4-b75b-b7ddf035402a-serving-cert\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.172146 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/49060193-d216-473c-8b67-6eb0d8970dab-bound-sa-token\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.176530 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mqjb"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.183752 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqf9p\" (UniqueName: \"kubernetes.io/projected/f2cf432c-1f44-4dd8-9f0b-c536db37b36b-kube-api-access-bqf9p\") pod \"migrator-59844c95c7-22cb7\" (UID: \"f2cf432c-1f44-4dd8-9f0b-c536db37b36b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.207027 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd4j8\" (UniqueName: \"kubernetes.io/projected/5271ef96-e4c6-49a4-b75b-b7ddf035402a-kube-api-access-jd4j8\") pod \"controller-manager-879f6c89f-tjzqr\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.207665 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-bound-sa-token\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.212362 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.240708 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.240962 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241047 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241076 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-mountpoint-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241107 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/577762e0-d3ab-4927-8561-fdd79b43f5b1-srv-cert\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241129 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a540a9ba-ecd7-4490-805b-45ff88d97a63-webhook-cert\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241149 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-proxy-tls\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241179 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241202 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf84026-ecfa-446d-91dc-e8146d350877-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241223 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15579146-527c-4348-9e39-6a9d0ba6533e-metrics-tls\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241241 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-config\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241270 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p4s9\" (UniqueName: \"kubernetes.io/projected/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-kube-api-access-4p4s9\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241291 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d34ff06-5475-4d4f-9e69-c0734d22554b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241314 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/73c91ec0-f08c-4a00-8b90-a8443d548128-certs\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5j6m\" (UniqueName: \"kubernetes.io/projected/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-kube-api-access-l5j6m\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241357 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxqp9\" (UniqueName: \"kubernetes.io/projected/c239d916-4d86-47a9-b2c5-cac73bcea6bf-kube-api-access-qxqp9\") pod \"package-server-manager-789f6589d5-6xlsv\" (UID: \"c239d916-4d86-47a9-b2c5-cac73bcea6bf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241393 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-csi-data-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241415 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-registration-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241435 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241466 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5496d724-4b33-4818-a317-d1ace3c64aeb-proxy-tls\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241496 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-signing-key\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a540a9ba-ecd7-4490-805b-45ff88d97a63-apiservice-cert\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241538 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55sc4\" (UniqueName: \"kubernetes.io/projected/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-kube-api-access-55sc4\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241564 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/309a1b03-85dc-466d-afdb-8c946fd030e1-serving-cert\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241582 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c5js\" (UniqueName: \"kubernetes.io/projected/309a1b03-85dc-466d-afdb-8c946fd030e1-kube-api-access-8c5js\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241597 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8czsm\" (UniqueName: \"kubernetes.io/projected/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-kube-api-access-8czsm\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241619 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/577762e0-d3ab-4927-8561-fdd79b43f5b1-profile-collector-cert\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241642 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/edd79356-997e-4b56-b884-921db4a45d38-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h6hqm\" (UID: \"edd79356-997e-4b56-b884-921db4a45d38\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241659 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/daf84026-ecfa-446d-91dc-e8146d350877-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241675 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66tb\" (UniqueName: \"kubernetes.io/projected/577762e0-d3ab-4927-8561-fdd79b43f5b1-kube-api-access-f66tb\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-signing-cabundle\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241713 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d34ff06-5475-4d4f-9e69-c0734d22554b-config\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241730 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9d4f\" (UniqueName: \"kubernetes.io/projected/15579146-527c-4348-9e39-6a9d0ba6533e-kube-api-access-t9d4f\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241747 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-images\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241765 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65s4q\" (UniqueName: \"kubernetes.io/projected/a540a9ba-ecd7-4490-805b-45ff88d97a63-kube-api-access-65s4q\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241782 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241820 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f63ad2d2-d528-48e8-99f8-891d91f72ded-metrics-tls\") pod \"dns-operator-744455d44c-mvv7h\" (UID: \"f63ad2d2-d528-48e8-99f8-891d91f72ded\") " pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241835 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j94d5\" (UniqueName: \"kubernetes.io/projected/5496d724-4b33-4818-a317-d1ace3c64aeb-kube-api-access-j94d5\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241857 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/be7f735b-ccce-4060-915d-58aea627d7f5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cwrnl\" (UID: \"be7f735b-ccce-4060-915d-58aea627d7f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241875 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/73c91ec0-f08c-4a00-8b90-a8443d548128-node-bootstrap-token\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241893 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/309a1b03-85dc-466d-afdb-8c946fd030e1-config\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241919 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhwjf\" (UniqueName: \"kubernetes.io/projected/73c91ec0-f08c-4a00-8b90-a8443d548128-kube-api-access-mhwjf\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241943 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g9v5\" (UniqueName: \"kubernetes.io/projected/f63ad2d2-d528-48e8-99f8-891d91f72ded-kube-api-access-9g9v5\") pod \"dns-operator-744455d44c-mvv7h\" (UID: \"f63ad2d2-d528-48e8-99f8-891d91f72ded\") " pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.241965 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wq4z\" (UniqueName: \"kubernetes.io/projected/a355257c-502e-495a-97f0-bc65d5f1cfd5-kube-api-access-4wq4z\") pod \"ingress-canary-j6mpg\" (UID: \"a355257c-502e-495a-97f0-bc65d5f1cfd5\") " pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242013 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-config-volume\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242032 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a355257c-502e-495a-97f0-bc65d5f1cfd5-cert\") pod \"ingress-canary-j6mpg\" (UID: \"a355257c-502e-495a-97f0-bc65d5f1cfd5\") " pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242054 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcrpg\" (UniqueName: \"kubernetes.io/projected/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-kube-api-access-pcrpg\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242075 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2p6v\" (UniqueName: \"kubernetes.io/projected/daf84026-ecfa-446d-91dc-e8146d350877-kube-api-access-n2p6v\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242297 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5jb5\" (UniqueName: \"kubernetes.io/projected/be7f735b-ccce-4060-915d-58aea627d7f5-kube-api-access-v5jb5\") pod \"control-plane-machine-set-operator-78cbb6b69f-cwrnl\" (UID: \"be7f735b-ccce-4060-915d-58aea627d7f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242319 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242334 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-secret-volume\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242350 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5496d724-4b33-4818-a317-d1ace3c64aeb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242365 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c239d916-4d86-47a9-b2c5-cac73bcea6bf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6xlsv\" (UID: \"c239d916-4d86-47a9-b2c5-cac73bcea6bf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242383 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-socket-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242400 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242424 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a540a9ba-ecd7-4490-805b-45ff88d97a63-tmpfs\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242440 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tplx\" (UniqueName: \"kubernetes.io/projected/edd79356-997e-4b56-b884-921db4a45d38-kube-api-access-5tplx\") pod \"multus-admission-controller-857f4d67dd-h6hqm\" (UID: \"edd79356-997e-4b56-b884-921db4a45d38\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242456 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d34ff06-5475-4d4f-9e69-c0734d22554b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242471 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-plugins-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.242488 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:22.742470091 +0000 UTC m=+94.755748618 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242528 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmjjh\" (UniqueName: \"kubernetes.io/projected/889e3661-2ba6-45ed-a371-3d9b6300db64-kube-api-access-tmjjh\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242565 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15579146-527c-4348-9e39-6a9d0ba6533e-config-volume\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.242627 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-plugins-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.243188 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15579146-527c-4348-9e39-6a9d0ba6533e-config-volume\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.246583 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-mountpoint-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.249547 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.251700 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/577762e0-d3ab-4927-8561-fdd79b43f5b1-profile-collector-cert\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.252153 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/edd79356-997e-4b56-b884-921db4a45d38-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h6hqm\" (UID: \"edd79356-997e-4b56-b884-921db4a45d38\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.252842 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-images\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.253017 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-signing-cabundle\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.254660 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-config-volume\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.255229 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d34ff06-5475-4d4f-9e69-c0734d22554b-config\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.256701 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-socket-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.257745 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a540a9ba-ecd7-4490-805b-45ff88d97a63-tmpfs\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.259693 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5496d724-4b33-4818-a317-d1ace3c64aeb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.260502 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f63ad2d2-d528-48e8-99f8-891d91f72ded-metrics-tls\") pod \"dns-operator-744455d44c-mvv7h\" (UID: \"f63ad2d2-d528-48e8-99f8-891d91f72ded\") " pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.261400 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a355257c-502e-495a-97f0-bc65d5f1cfd5-cert\") pod \"ingress-canary-j6mpg\" (UID: \"a355257c-502e-495a-97f0-bc65d5f1cfd5\") " pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.261912 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/be7f735b-ccce-4060-915d-58aea627d7f5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cwrnl\" (UID: \"be7f735b-ccce-4060-915d-58aea627d7f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.262531 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/daf84026-ecfa-446d-91dc-e8146d350877-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.262793 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a540a9ba-ecd7-4490-805b-45ff88d97a63-webhook-cert\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.263220 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c239d916-4d86-47a9-b2c5-cac73bcea6bf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6xlsv\" (UID: \"c239d916-4d86-47a9-b2c5-cac73bcea6bf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.263846 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-csi-data-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.264028 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/889e3661-2ba6-45ed-a371-3d9b6300db64-registration-dir\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.264066 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.264641 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/309a1b03-85dc-466d-afdb-8c946fd030e1-config\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.265365 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.265409 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.267117 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv7nz\" (UniqueName: \"kubernetes.io/projected/7adf9170-1515-4c07-b471-bcc775bf4327-kube-api-access-cv7nz\") pod \"openshift-config-operator-7777fb866f-2r2gh\" (UID: \"7adf9170-1515-4c07-b471-bcc775bf4327\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.267532 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-config\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.268107 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf84026-ecfa-446d-91dc-e8146d350877-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.270098 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15579146-527c-4348-9e39-6a9d0ba6533e-metrics-tls\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.271052 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.272644 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.280453 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-secret-volume\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.280588 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/577762e0-d3ab-4927-8561-fdd79b43f5b1-srv-cert\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.281169 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.283325 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.284013 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-signing-key\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.283610 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/73c91ec0-f08c-4a00-8b90-a8443d548128-node-bootstrap-token\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.284691 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/73c91ec0-f08c-4a00-8b90-a8443d548128-certs\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.285565 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5496d724-4b33-4818-a317-d1ace3c64aeb-proxy-tls\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.285928 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a540a9ba-ecd7-4490-805b-45ff88d97a63-apiservice-cert\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.286432 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d34ff06-5475-4d4f-9e69-c0734d22554b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.287162 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/309a1b03-85dc-466d-afdb-8c946fd030e1-serving-cert\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.288998 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml24h\" (UniqueName: \"kubernetes.io/projected/ecf5a09b-2113-4202-83c1-ad8c77520791-kube-api-access-ml24h\") pod \"machine-api-operator-5694c8668f-pbnlg\" (UID: \"ecf5a09b-2113-4202-83c1-ad8c77520791\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.289160 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.292484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-proxy-tls\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.303504 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzt5s\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-kube-api-access-xzt5s\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.321856 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5sjs\" (UniqueName: \"kubernetes.io/projected/47d764ec-58e3-4604-a3b6-259dc3c3fa36-kube-api-access-j5sjs\") pod \"apiserver-7bbb656c7d-gc8lg\" (UID: \"47d764ec-58e3-4604-a3b6-259dc3c3fa36\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.324457 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czkqh\" (UniqueName: \"kubernetes.io/projected/49060193-d216-473c-8b67-6eb0d8970dab-kube-api-access-czkqh\") pod \"ingress-operator-5b745b69d9-c9p6v\" (UID: \"49060193-d216-473c-8b67-6eb0d8970dab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.325090 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6r92c" event={"ID":"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2","Type":"ContainerStarted","Data":"9ad25bc125eb3284dd3b651262d76a4bdd79492d86923a4366f2f28fbbe55815"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.330011 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h5scx" event={"ID":"73dd5dd2-4416-4804-9377-e66521991a5e","Type":"ContainerStarted","Data":"30c047a7f843f7a41fd44c82b554c1e6d8288e007cf65fc744e0f7b12e711e06"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.330326 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h5scx" event={"ID":"73dd5dd2-4416-4804-9377-e66521991a5e","Type":"ContainerStarted","Data":"2604e28d731428acfc5465562e5d6a058c375e33ca39e88e018651579db10b5e"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.330943 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.331841 4796 patch_prober.go:28] interesting pod/console-operator-58897d9998-h5scx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.331874 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-h5scx" podUID="73dd5dd2-4416-4804-9377-e66521991a5e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.333897 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" event={"ID":"8adba875-c47a-49d0-a52e-5ddc97bfb928","Type":"ContainerStarted","Data":"2ebad35a74822671b461225f2305b11c785b19edf8abd5e65307e369baf2ce87"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.333922 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" event={"ID":"8adba875-c47a-49d0-a52e-5ddc97bfb928","Type":"ContainerStarted","Data":"fc78d51d8eb4f11ac53af4c33b819acd78b68102140370acd68bdf114488db12"} Sep 30 16:13:22 crc kubenswrapper[4796]: W0930 16:13:22.337685 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89bad259_b74e_4bd2_b8cd_64b59ccc6ae6.slice/crio-50d5e64161bac8ba3bdac4d599d27fc12a49edc371dbc589f930aa5851654033 WatchSource:0}: Error finding container 50d5e64161bac8ba3bdac4d599d27fc12a49edc371dbc589f930aa5851654033: Status 404 returned error can't find the container with id 50d5e64161bac8ba3bdac4d599d27fc12a49edc371dbc589f930aa5851654033 Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.339438 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" event={"ID":"73414d05-f9c7-49fc-9770-d12886f12382","Type":"ContainerStarted","Data":"f836541722646f657835d9f50cd4918c9838a3c535e7c5aa9add783620dea0e5"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.345235 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.346409 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:22.84639407 +0000 UTC m=+94.859672597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.347167 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.355140 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" event={"ID":"839c7aeb-ed29-4196-9fc5-0a8eecf7e585","Type":"ContainerStarted","Data":"a7917a9d1d0b885a337debc48ee1d51a6b6190a8b1f4d7de5625e337b9bf4e45"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.355187 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" event={"ID":"839c7aeb-ed29-4196-9fc5-0a8eecf7e585","Type":"ContainerStarted","Data":"f95927b308588471faee616c7802a27cb9cea2aaaf9f44606ecbce7417abf553"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.356358 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.357510 4796 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gp486 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.357547 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" podUID="839c7aeb-ed29-4196-9fc5-0a8eecf7e585" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.366843 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" event={"ID":"2434945c-60ef-40c1-aa33-2cded03cc0e0","Type":"ContainerStarted","Data":"694dba313a4a204045eb8dd2ee43659fd696a153c497e83a99d0ad93fec3d0b9"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.367020 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c5js\" (UniqueName: \"kubernetes.io/projected/309a1b03-85dc-466d-afdb-8c946fd030e1-kube-api-access-8c5js\") pod \"service-ca-operator-777779d784-xm9vj\" (UID: \"309a1b03-85dc-466d-afdb-8c946fd030e1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.369087 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" event={"ID":"290a8e80-8ade-4087-84f3-8223e380c0ef","Type":"ContainerStarted","Data":"986e96b471805c812a08d53417a45da4dc4dbfa8b134bbd56a7c89b6878aae8e"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.369107 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" event={"ID":"290a8e80-8ade-4087-84f3-8223e380c0ef","Type":"ContainerStarted","Data":"87abb0c44f3b9e333dd3df59734b06a10b503a1afec1c8c020bc90b22abf619f"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.373341 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ggvd5"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.373366 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" event={"ID":"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f","Type":"ContainerStarted","Data":"c45dc4c3938d5c355a8af871959128cda8f0a6088e51684036fe09d1484db3bd"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.373382 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" event={"ID":"6b8528d4-f5f0-4d0a-b7a6-2d283ef9d62f","Type":"ContainerStarted","Data":"34e794d2503c3dd87e1c374d8b13b06f8b77f61f2b4bcdcb1102137c6ccd1ed0"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.376705 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" event={"ID":"06101703-e6d8-42e5-8549-7340fbf1a597","Type":"ContainerStarted","Data":"0d0a59ade7f0b317e112014e8c20f06162da7e4add2b1c1456371dc332f75a6c"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.378389 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" event={"ID":"4c6f0a6e-14ff-45a7-8312-09fd13121728","Type":"ContainerStarted","Data":"2cced5da2bcdc9c69901de78b02df456841aab910c08efd4c6389bc8b89a394c"} Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.387966 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8czsm\" (UniqueName: \"kubernetes.io/projected/cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc-kube-api-access-8czsm\") pod \"machine-config-operator-74547568cd-mzcjx\" (UID: \"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.399060 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66tb\" (UniqueName: \"kubernetes.io/projected/577762e0-d3ab-4927-8561-fdd79b43f5b1-kube-api-access-f66tb\") pod \"catalog-operator-68c6474976-cps74\" (UID: \"577762e0-d3ab-4927-8561-fdd79b43f5b1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.411792 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.422664 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9d4f\" (UniqueName: \"kubernetes.io/projected/15579146-527c-4348-9e39-6a9d0ba6533e-kube-api-access-t9d4f\") pod \"dns-default-mlhmv\" (UID: \"15579146-527c-4348-9e39-6a9d0ba6533e\") " pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.442820 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dxc8p"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.443040 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9f8bd48-98d2-469a-ad2b-37ad58c7a23d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fsrld\" (UID: \"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.448079 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.449051 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:22.949037282 +0000 UTC m=+94.962315809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: W0930 16:13:22.450467 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b2cdffa_2549_4fb3_aa14_6b2da0738146.slice/crio-a6c60499fe3bfe8c9847794404a11fc77bc9ce761178224ec943505b8cf8f9a9 WatchSource:0}: Error finding container a6c60499fe3bfe8c9847794404a11fc77bc9ce761178224ec943505b8cf8f9a9: Status 404 returned error can't find the container with id a6c60499fe3bfe8c9847794404a11fc77bc9ce761178224ec943505b8cf8f9a9 Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.452662 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.457007 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.469738 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjzqr"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.477044 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.478203 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65s4q\" (UniqueName: \"kubernetes.io/projected/a540a9ba-ecd7-4490-805b-45ff88d97a63-kube-api-access-65s4q\") pod \"packageserver-d55dfcdfc-h49fq\" (UID: \"a540a9ba-ecd7-4490-805b-45ff88d97a63\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.484873 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.488119 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g9v5\" (UniqueName: \"kubernetes.io/projected/f63ad2d2-d528-48e8-99f8-891d91f72ded-kube-api-access-9g9v5\") pod \"dns-operator-744455d44c-mvv7h\" (UID: \"f63ad2d2-d528-48e8-99f8-891d91f72ded\") " pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: W0930 16:13:22.492973 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5271ef96_e4c6_49a4_b75b_b7ddf035402a.slice/crio-c11c76bb71f026108901068306d3e72c2b3484cc21a3982fb31d896f018944f4 WatchSource:0}: Error finding container c11c76bb71f026108901068306d3e72c2b3484cc21a3982fb31d896f018944f4: Status 404 returned error can't find the container with id c11c76bb71f026108901068306d3e72c2b3484cc21a3982fb31d896f018944f4 Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.501340 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wq4z\" (UniqueName: \"kubernetes.io/projected/a355257c-502e-495a-97f0-bc65d5f1cfd5-kube-api-access-4wq4z\") pod \"ingress-canary-j6mpg\" (UID: \"a355257c-502e-495a-97f0-bc65d5f1cfd5\") " pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.504660 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cch5s"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.511634 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.522426 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcrpg\" (UniqueName: \"kubernetes.io/projected/d94cc59a-e21c-45f5-a1d2-0b82fed85e21-kube-api-access-pcrpg\") pod \"openshift-controller-manager-operator-756b6f6bc6-hc9z9\" (UID: \"d94cc59a-e21c-45f5-a1d2-0b82fed85e21\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.522572 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.537175 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-j6mpg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.539490 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.545411 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2p6v\" (UniqueName: \"kubernetes.io/projected/daf84026-ecfa-446d-91dc-e8146d350877-kube-api-access-n2p6v\") pod \"kube-storage-version-migrator-operator-b67b599dd-pg2pz\" (UID: \"daf84026-ecfa-446d-91dc-e8146d350877\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.549162 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.549450 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.04943976 +0000 UTC m=+95.062718287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.551167 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.564779 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5jb5\" (UniqueName: \"kubernetes.io/projected/be7f735b-ccce-4060-915d-58aea627d7f5-kube-api-access-v5jb5\") pod \"control-plane-machine-set-operator-78cbb6b69f-cwrnl\" (UID: \"be7f735b-ccce-4060-915d-58aea627d7f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.589551 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j94d5\" (UniqueName: \"kubernetes.io/projected/5496d724-4b33-4818-a317-d1ace3c64aeb-kube-api-access-j94d5\") pod \"machine-config-controller-84d6567774-tzrf4\" (UID: \"5496d724-4b33-4818-a317-d1ace3c64aeb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.604589 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.624454 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tplx\" (UniqueName: \"kubernetes.io/projected/edd79356-997e-4b56-b884-921db4a45d38-kube-api-access-5tplx\") pod \"multus-admission-controller-857f4d67dd-h6hqm\" (UID: \"edd79356-997e-4b56-b884-921db4a45d38\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.624696 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d34ff06-5475-4d4f-9e69-c0734d22554b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-m8lpk\" (UID: \"8d34ff06-5475-4d4f-9e69-c0734d22554b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.644554 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhwjf\" (UniqueName: \"kubernetes.io/projected/73c91ec0-f08c-4a00-8b90-a8443d548128-kube-api-access-mhwjf\") pod \"machine-config-server-sl6k6\" (UID: \"73c91ec0-f08c-4a00-8b90-a8443d548128\") " pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.651827 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.652056 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.152027711 +0000 UTC m=+95.165306238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.652199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.652492 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.152480284 +0000 UTC m=+95.165758811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: W0930 16:13:22.663754 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7adf9170_1515_4c07_b471_bcc775bf4327.slice/crio-ab3a98982aa944f69941691ee90b60ee3ddc39941bcc16a2880712f49094bc0b WatchSource:0}: Error finding container ab3a98982aa944f69941691ee90b60ee3ddc39941bcc16a2880712f49094bc0b: Status 404 returned error can't find the container with id ab3a98982aa944f69941691ee90b60ee3ddc39941bcc16a2880712f49094bc0b Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.672102 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxqp9\" (UniqueName: \"kubernetes.io/projected/c239d916-4d86-47a9-b2c5-cac73bcea6bf-kube-api-access-qxqp9\") pod \"package-server-manager-789f6589d5-6xlsv\" (UID: \"c239d916-4d86-47a9-b2c5-cac73bcea6bf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.683455 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55sc4\" (UniqueName: \"kubernetes.io/projected/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-kube-api-access-55sc4\") pod \"marketplace-operator-79b997595-bbg46\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.707834 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p4s9\" (UniqueName: \"kubernetes.io/projected/3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7-kube-api-access-4p4s9\") pod \"service-ca-9c57cc56f-s9qpw\" (UID: \"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7\") " pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.720746 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.729800 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.732511 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmjjh\" (UniqueName: \"kubernetes.io/projected/889e3661-2ba6-45ed-a371-3d9b6300db64-kube-api-access-tmjjh\") pod \"csi-hostpathplugin-fw6t6\" (UID: \"889e3661-2ba6-45ed-a371-3d9b6300db64\") " pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.736012 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.745374 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.747451 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5j6m\" (UniqueName: \"kubernetes.io/projected/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-kube-api-access-l5j6m\") pod \"collect-profiles-29320800-pdt4z\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.751883 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.753721 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.753857 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.25383546 +0000 UTC m=+95.267113987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.753999 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.754301 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.254292973 +0000 UTC m=+95.267571500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.767122 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.771780 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.780002 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.796255 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.801768 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.811228 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.816868 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.829889 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.847695 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sl6k6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.856724 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.857084 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.357069169 +0000 UTC m=+95.370347696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:22 crc kubenswrapper[4796]: W0930 16:13:22.858084 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2cf432c_1f44_4dd8_9f0b_c536db37b36b.slice/crio-ac25c83d1bc2debdce407a7ce2b62c91e811fb399407e6510c72e59fc981870b WatchSource:0}: Error finding container ac25c83d1bc2debdce407a7ce2b62c91e811fb399407e6510c72e59fc981870b: Status 404 returned error can't find the container with id ac25c83d1bc2debdce407a7ce2b62c91e811fb399407e6510c72e59fc981870b Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.873937 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.903162 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.922724 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj"] Sep 30 16:13:22 crc kubenswrapper[4796]: I0930 16:13:22.958604 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:22 crc kubenswrapper[4796]: E0930 16:13:22.959041 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.459022131 +0000 UTC m=+95.472300648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.042898 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.059478 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.059729 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.559705348 +0000 UTC m=+95.572983875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.059954 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.060389 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.560381607 +0000 UTC m=+95.573660124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.161135 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.161493 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.661478345 +0000 UTC m=+95.674756862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: W0930 16:13:23.182046 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73c91ec0_f08c_4a00_8b90_a8443d548128.slice/crio-677de40e61c7962f50a2acf87091b70c872cdfa6d58888586b2fac25791c73b5 WatchSource:0}: Error finding container 677de40e61c7962f50a2acf87091b70c872cdfa6d58888586b2fac25791c73b5: Status 404 returned error can't find the container with id 677de40e61c7962f50a2acf87091b70c872cdfa6d58888586b2fac25791c73b5 Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.263215 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.263878 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.76386131 +0000 UTC m=+95.777139837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.294937 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pbnlg"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.316632 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.365113 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.365844 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.865823483 +0000 UTC m=+95.879102010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.385949 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.404158 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.430520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sl6k6" event={"ID":"73c91ec0-f08c-4a00-8b90-a8443d548128","Type":"ContainerStarted","Data":"677de40e61c7962f50a2acf87091b70c872cdfa6d58888586b2fac25791c73b5"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.435255 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" event={"ID":"9f36e801-bfb5-425f-876d-ed8191d5e66e","Type":"ContainerStarted","Data":"61c87897b596ec6a70132009859f1c810e01aeaacf1acfd335acf7ba4b74016a"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.435294 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" event={"ID":"9f36e801-bfb5-425f-876d-ed8191d5e66e","Type":"ContainerStarted","Data":"0707067110a47058083ba2180bfdaed082cd8c8f0c014c997152d6c3da9e4915"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.438973 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" event={"ID":"73414d05-f9c7-49fc-9770-d12886f12382","Type":"ContainerStarted","Data":"2228df7d12cf6b905c958bd5481d0f805d17f18a79c3bea75695357df543ac20"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.440465 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" event={"ID":"49060193-d216-473c-8b67-6eb0d8970dab","Type":"ContainerStarted","Data":"edd9fca18b0ed7eb7f86a1b574cf86527339bcdf204a4eabea6434eb86888605"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.441666 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dxc8p" event={"ID":"025a66cf-d82d-4590-b9cd-402bc6bc3668","Type":"ContainerStarted","Data":"843a998c1828d01b10c484614d9a478f31b68cab49a6fdddf73f39b8fa977915"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.441792 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dxc8p" event={"ID":"025a66cf-d82d-4590-b9cd-402bc6bc3668","Type":"ContainerStarted","Data":"fb83ac620693eb904e2ce1855a1ea1550b13d4aa84f7f5f9fdd521a437687776"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.441966 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.443046 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" event={"ID":"2434945c-60ef-40c1-aa33-2cded03cc0e0","Type":"ContainerStarted","Data":"deae6dfad6e2e288357909a067910f107093a549fc8925bc1f521a80aa440162"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.443248 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.443819 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-dxc8p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.443854 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dxc8p" podUID="025a66cf-d82d-4590-b9cd-402bc6bc3668" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.446603 4796 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9mqjb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.446669 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" podUID="2434945c-60ef-40c1-aa33-2cded03cc0e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.447079 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6r92c" event={"ID":"fb586c4b-79c4-4747-b9d4-8ff7e21e03f2","Type":"ContainerStarted","Data":"21020dc7920dd0ce860a3920fb856fe0f3540740086f3fc4e182c868a5996a25"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.458331 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" event={"ID":"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc","Type":"ContainerStarted","Data":"6b122c37cd434a3fc30c586571d86f6891712bc93501f3a67f4435de367e3238"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.469087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.469828 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:23.969812303 +0000 UTC m=+95.983090830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.473043 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mlhmv"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.485913 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-j6mpg"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.487082 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" event={"ID":"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6","Type":"ContainerStarted","Data":"577a42dcb7c824000b52c2a69168d07b54bc585e5f27b9683c37a785afb67e98"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.487112 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" event={"ID":"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6","Type":"ContainerStarted","Data":"50d5e64161bac8ba3bdac4d599d27fc12a49edc371dbc589f930aa5851654033"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.487128 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.496489 4796 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-6dvjn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.496829 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" podUID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.510787 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cch5s" event={"ID":"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80","Type":"ContainerStarted","Data":"990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.510837 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cch5s" event={"ID":"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80","Type":"ContainerStarted","Data":"99c5a05428eae90dafbc3c3ab185d699edbcf71df1b1b5a1f08fa48b08c4584e"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.525713 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" event={"ID":"0baddf30-b72f-4659-b221-400c7b11ad0c","Type":"ContainerStarted","Data":"8e0cf80099f9a92ec12c85c3f044a95d89984f25c29fe7d6854f0eb07ced0450"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.525759 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" event={"ID":"0baddf30-b72f-4659-b221-400c7b11ad0c","Type":"ContainerStarted","Data":"9380c708f76e4a91dd13ad3764b33bfea74714fe43c04108d7f02279952312f4"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.539597 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" event={"ID":"4c6f0a6e-14ff-45a7-8312-09fd13121728","Type":"ContainerStarted","Data":"7e3b605662bd9f6920dc32f8c8e6953f074aa850345d0630b78a9fa4d2cc59c5"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.551444 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" event={"ID":"7adf9170-1515-4c07-b471-bcc775bf4327","Type":"ContainerStarted","Data":"542360c40490203007e085bb4278caed4be2080a17772b20541bdb1fce8fd94d"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.551495 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" event={"ID":"7adf9170-1515-4c07-b471-bcc775bf4327","Type":"ContainerStarted","Data":"ab3a98982aa944f69941691ee90b60ee3ddc39941bcc16a2880712f49094bc0b"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.570296 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.576933 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.076911443 +0000 UTC m=+96.090189970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.577609 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" event={"ID":"309a1b03-85dc-466d-afdb-8c946fd030e1","Type":"ContainerStarted","Data":"50b65188dcedb8234dec47aba2f2a4728ab35b89ea1544a9aa8af314878cdb25"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.597018 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" event={"ID":"1b2cdffa-2549-4fb3-aa14-6b2da0738146","Type":"ContainerStarted","Data":"ae081e64f9b5c1c985f2498628270b13e6ced04490f4daa67f173d083f04a9d8"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.597068 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" event={"ID":"1b2cdffa-2549-4fb3-aa14-6b2da0738146","Type":"ContainerStarted","Data":"a6c60499fe3bfe8c9847794404a11fc77bc9ce761178224ec943505b8cf8f9a9"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.624839 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v222h" podStartSLOduration=69.624803541 podStartE2EDuration="1m9.624803541s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:23.578629682 +0000 UTC m=+95.591908209" watchObservedRunningTime="2025-09-30 16:13:23.624803541 +0000 UTC m=+95.638082068" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.625341 4796 generic.go:334] "Generic (PLEG): container finished" podID="06101703-e6d8-42e5-8549-7340fbf1a597" containerID="5367b5cd01edc272a075b0f418351d175890989748d5f344370d3a2ebf3513ca" exitCode=0 Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.625660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" event={"ID":"06101703-e6d8-42e5-8549-7340fbf1a597","Type":"ContainerDied","Data":"5367b5cd01edc272a075b0f418351d175890989748d5f344370d3a2ebf3513ca"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.634214 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mvv7h"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.640366 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.645465 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq"] Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.647193 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sj2sq" podStartSLOduration=69.647155329 podStartE2EDuration="1m9.647155329s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:23.621480816 +0000 UTC m=+95.634759343" watchObservedRunningTime="2025-09-30 16:13:23.647155329 +0000 UTC m=+95.660433856" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.652862 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" event={"ID":"f2cf432c-1f44-4dd8-9f0b-c536db37b36b","Type":"ContainerStarted","Data":"ac25c83d1bc2debdce407a7ce2b62c91e811fb399407e6510c72e59fc981870b"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.662766 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" event={"ID":"5271ef96-e4c6-49a4-b75b-b7ddf035402a","Type":"ContainerStarted","Data":"312c8723b3ee39245f8973360e13912b5368bd365df91ea6e13ed34bd0d453c4"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.662849 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.662865 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" event={"ID":"5271ef96-e4c6-49a4-b75b-b7ddf035402a","Type":"ContainerStarted","Data":"c11c76bb71f026108901068306d3e72c2b3484cc21a3982fb31d896f018944f4"} Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.665111 4796 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tjzqr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.665167 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" podUID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.668577 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.674044 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.675781 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.175766827 +0000 UTC m=+96.189045354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.737609 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6svgm" podStartSLOduration=69.737589673 podStartE2EDuration="1m9.737589673s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:23.70246745 +0000 UTC m=+95.715745977" watchObservedRunningTime="2025-09-30 16:13:23.737589673 +0000 UTC m=+95.750868200" Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.777721 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.791608 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.291573845 +0000 UTC m=+96.304852372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: W0930 16:13:23.791713 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda355257c_502e_495a_97f0_bc65d5f1cfd5.slice/crio-ebd9d065efe998dbef548de92d73089ed22f890152b47e967082bcfdd7ea6e24 WatchSource:0}: Error finding container ebd9d065efe998dbef548de92d73089ed22f890152b47e967082bcfdd7ea6e24: Status 404 returned error can't find the container with id ebd9d065efe998dbef548de92d73089ed22f890152b47e967082bcfdd7ea6e24 Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.879917 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.880596 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.380585388 +0000 UTC m=+96.393863915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:23 crc kubenswrapper[4796]: W0930 16:13:23.929375 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf63ad2d2_d528_48e8_99f8_891d91f72ded.slice/crio-b18f34d2b03e286c60f0cbd823f50d4a6cc06cf389e5ae17617f68b7bfa3a516 WatchSource:0}: Error finding container b18f34d2b03e286c60f0cbd823f50d4a6cc06cf389e5ae17617f68b7bfa3a516: Status 404 returned error can't find the container with id b18f34d2b03e286c60f0cbd823f50d4a6cc06cf389e5ae17617f68b7bfa3a516 Sep 30 16:13:23 crc kubenswrapper[4796]: I0930 16:13:23.981150 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:23 crc kubenswrapper[4796]: E0930 16:13:23.981478 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.48146039 +0000 UTC m=+96.494738917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.041113 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gp486" podStartSLOduration=69.041092693 podStartE2EDuration="1m9.041092693s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:23.990254431 +0000 UTC m=+96.003532958" watchObservedRunningTime="2025-09-30 16:13:24.041092693 +0000 UTC m=+96.054371220" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.073645 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-h5scx" podStartSLOduration=70.073626293 podStartE2EDuration="1m10.073626293s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.043125671 +0000 UTC m=+96.056404188" watchObservedRunningTime="2025-09-30 16:13:24.073626293 +0000 UTC m=+96.086904820" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.077725 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.088410 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.088689 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.588678423 +0000 UTC m=+96.601956950 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.099234 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:24 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:24 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:24 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.099283 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.100011 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.152919 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbg46"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.157308 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.189240 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.189746 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.689726299 +0000 UTC m=+96.703004826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.328204 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.328564 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.828550255 +0000 UTC m=+96.841828782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.335204 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h6hqm"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.349934 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.397334 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-h5scx" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.424507 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dxc8p" podStartSLOduration=70.424491086 podStartE2EDuration="1m10.424491086s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.423723044 +0000 UTC m=+96.437001571" watchObservedRunningTime="2025-09-30 16:13:24.424491086 +0000 UTC m=+96.437769613" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.430764 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.431146 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:24.931134426 +0000 UTC m=+96.944412953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.510610 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-cch5s" podStartSLOduration=70.510596355 podStartE2EDuration="1m10.510596355s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.509243326 +0000 UTC m=+96.522521853" watchObservedRunningTime="2025-09-30 16:13:24.510596355 +0000 UTC m=+96.523874882" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.511688 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l8rrg" podStartSLOduration=69.511683256 podStartE2EDuration="1m9.511683256s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.463793899 +0000 UTC m=+96.477072426" watchObservedRunningTime="2025-09-30 16:13:24.511683256 +0000 UTC m=+96.524961773" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.517740 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fw6t6"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.532653 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.533027 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.033009775 +0000 UTC m=+97.046288302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.540322 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.545112 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" podStartSLOduration=69.54509872 podStartE2EDuration="1m9.54509872s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.543552586 +0000 UTC m=+96.556831113" watchObservedRunningTime="2025-09-30 16:13:24.54509872 +0000 UTC m=+96.558377247" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.555959 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.603853 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s9qpw"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.613749 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z"] Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.634469 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.634829 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.134813973 +0000 UTC m=+97.148092500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: W0930 16:13:24.655005 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dc25d62_cd1c_45aa_a1c9_740d9e9a5fa7.slice/crio-ac1a862a79b8ad6e7e40283254434e7fc495dba63a3a39801479ac540093401d WatchSource:0}: Error finding container ac1a862a79b8ad6e7e40283254434e7fc495dba63a3a39801479ac540093401d: Status 404 returned error can't find the container with id ac1a862a79b8ad6e7e40283254434e7fc495dba63a3a39801479ac540093401d Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.667800 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-6r92c" podStartSLOduration=69.667782575 podStartE2EDuration="1m9.667782575s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.630286774 +0000 UTC m=+96.643565301" watchObservedRunningTime="2025-09-30 16:13:24.667782575 +0000 UTC m=+96.681061102" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.668485 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" podStartSLOduration=70.668481345 podStartE2EDuration="1m10.668481345s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.666283462 +0000 UTC m=+96.679561979" watchObservedRunningTime="2025-09-30 16:13:24.668481345 +0000 UTC m=+96.681759872" Sep 30 16:13:24 crc kubenswrapper[4796]: W0930 16:13:24.683390 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod889e3661_2ba6_45ed_a371_3d9b6300db64.slice/crio-b29056ec4fe6a4c7f5e3796cb0e6db59f91e3c48ac694e22af1baef9460c4959 WatchSource:0}: Error finding container b29056ec4fe6a4c7f5e3796cb0e6db59f91e3c48ac694e22af1baef9460c4959: Status 404 returned error can't find the container with id b29056ec4fe6a4c7f5e3796cb0e6db59f91e3c48ac694e22af1baef9460c4959 Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.684015 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mlhmv" event={"ID":"15579146-527c-4348-9e39-6a9d0ba6533e","Type":"ContainerStarted","Data":"2dded94b7b45e5371714e2e6233f093a308e95ac33e85abce1d503b3d7404ccd"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.687023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" event={"ID":"577762e0-d3ab-4927-8561-fdd79b43f5b1","Type":"ContainerStarted","Data":"891bf3a8932dbf1849877b208736a4cdece63494be56b950131fc7fabe3b7931"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.687089 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.687111 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" event={"ID":"577762e0-d3ab-4927-8561-fdd79b43f5b1","Type":"ContainerStarted","Data":"0911ea6e44b7b6d6164481e9a95416323dff1bf41f2453ecffdf9ae27e4bc665"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.689223 4796 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cps74 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.689272 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" podUID="577762e0-d3ab-4927-8561-fdd79b43f5b1" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.689652 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" event={"ID":"edd79356-997e-4b56-b884-921db4a45d38","Type":"ContainerStarted","Data":"d046663cc39fe6511aa817d761bad22109f73fefc25b6d35ffc27ee04928eb51"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.692072 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" event={"ID":"5496d724-4b33-4818-a317-d1ace3c64aeb","Type":"ContainerStarted","Data":"763776bd367aa1a43331179e476c763c0907034f8a852e04ce28658bc57369bd"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.698910 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" event={"ID":"0baddf30-b72f-4659-b221-400c7b11ad0c","Type":"ContainerStarted","Data":"5474b59919708f834d1e7559941e3f9563ab59757a8f4667d441634d954540fe"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.700014 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-j6mpg" event={"ID":"a355257c-502e-495a-97f0-bc65d5f1cfd5","Type":"ContainerStarted","Data":"ebd9d065efe998dbef548de92d73089ed22f890152b47e967082bcfdd7ea6e24"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.716453 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" podStartSLOduration=70.716432905 podStartE2EDuration="1m10.716432905s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.700632784 +0000 UTC m=+96.713911321" watchObservedRunningTime="2025-09-30 16:13:24.716432905 +0000 UTC m=+96.729711432" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.736261 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.736712 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.236691944 +0000 UTC m=+97.249970471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.787627 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" event={"ID":"06101703-e6d8-42e5-8549-7340fbf1a597","Type":"ContainerStarted","Data":"49adba50f89fcf1bea888cd11041dba6adcf09754f1ba77bd5c69873c47e5910"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.788630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" event={"ID":"a540a9ba-ecd7-4490-805b-45ff88d97a63","Type":"ContainerStarted","Data":"5a2e2c081c114b6e3d5f3be1fb60842af88c6d4a0e6fa754cbf6ace64a6ef98b"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.794621 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8cfhw" podStartSLOduration=70.794589478 podStartE2EDuration="1m10.794589478s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.749291434 +0000 UTC m=+96.762569971" watchObservedRunningTime="2025-09-30 16:13:24.794589478 +0000 UTC m=+96.807868005" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.795416 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-ggvd5" podStartSLOduration=70.795409621 podStartE2EDuration="1m10.795409621s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.792517329 +0000 UTC m=+96.805795856" watchObservedRunningTime="2025-09-30 16:13:24.795409621 +0000 UTC m=+96.808688148" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.802894 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" event={"ID":"d94cc59a-e21c-45f5-a1d2-0b82fed85e21","Type":"ContainerStarted","Data":"aa6813675ba9e990587b0100018868253f28487f897732fcc047cae564f39450"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.802931 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" event={"ID":"d94cc59a-e21c-45f5-a1d2-0b82fed85e21","Type":"ContainerStarted","Data":"beaa1b89a93693234a1e952e69eb9fe42b55a5cab0aecc273667c4f3b118fa28"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.808890 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" event={"ID":"7881eaa3-77bc-430c-a2a6-d44ada95cc0a","Type":"ContainerStarted","Data":"fa77db16d03c19e95957e4fe8ab4d72dbefaf8b99a5b40548b7d3dd036f0739b"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.832807 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" event={"ID":"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d","Type":"ContainerStarted","Data":"633ca17367f02e302cd8c559729bcfc1be4bd7c5493cce61f9eb1bb1bc676a87"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.837061 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.841059 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.341039315 +0000 UTC m=+97.354317842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.861011 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hc9z9" podStartSLOduration=70.860995565 podStartE2EDuration="1m10.860995565s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.860423228 +0000 UTC m=+96.873701755" watchObservedRunningTime="2025-09-30 16:13:24.860995565 +0000 UTC m=+96.874274092" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.867938 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" event={"ID":"f2cf432c-1f44-4dd8-9f0b-c536db37b36b","Type":"ContainerStarted","Data":"aaa432aa3fae06e057f7c4289156f332dc0d3d7260d7c762252347b8b98376be"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.867994 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" event={"ID":"f2cf432c-1f44-4dd8-9f0b-c536db37b36b","Type":"ContainerStarted","Data":"818b7dd9e0dac69e9ac19a232df1982b7ed20e021182e2e10ea64184a3be368f"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.888847 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" event={"ID":"47d764ec-58e3-4604-a3b6-259dc3c3fa36","Type":"ContainerStarted","Data":"5db2a9618b76b94aba9da949b8af41468fc75d66b44a581f6e7999185fa8d3a9"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.906876 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" event={"ID":"daf84026-ecfa-446d-91dc-e8146d350877","Type":"ContainerStarted","Data":"43feefd96b9dcc8d441f2c57278c519172dffd575d5668c3094cd0608ddefcd5"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.916257 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" event={"ID":"be7f735b-ccce-4060-915d-58aea627d7f5","Type":"ContainerStarted","Data":"cd25b94b3ecd62045e0d29b9acbb2adc59ba792ec5a78a309fd563b7923e91a9"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.922901 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sl6k6" event={"ID":"73c91ec0-f08c-4a00-8b90-a8443d548128","Type":"ContainerStarted","Data":"40ffc668691aa2d805f89ac21dadf481c534dcbd30d688c920b893f511f6b32b"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.929443 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" event={"ID":"ecf5a09b-2113-4202-83c1-ad8c77520791","Type":"ContainerStarted","Data":"d814457ca0d5478a39efdfd2b663ea1bc383b2a66b04e1e0012b3ce7e086e386"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.934031 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" event={"ID":"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc","Type":"ContainerStarted","Data":"cb9bd0c702d532c8b9e6fca9bbf6cab09cf0a2a94e48c058138534d25ca15f5a"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.938616 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" event={"ID":"f63ad2d2-d528-48e8-99f8-891d91f72ded","Type":"ContainerStarted","Data":"b18f34d2b03e286c60f0cbd823f50d4a6cc06cf389e5ae17617f68b7bfa3a516"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.939278 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.939953 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rk8v2" podStartSLOduration=70.93994326 podStartE2EDuration="1m10.93994326s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.937727187 +0000 UTC m=+96.951005704" watchObservedRunningTime="2025-09-30 16:13:24.93994326 +0000 UTC m=+96.953221787" Sep 30 16:13:24 crc kubenswrapper[4796]: E0930 16:13:24.940305 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.4402924 +0000 UTC m=+97.453570927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.964644 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" event={"ID":"8d34ff06-5475-4d4f-9e69-c0734d22554b","Type":"ContainerStarted","Data":"b6b6b10ae4998897242e9210c290252847dce40678d89b3130369ec992f335bb"} Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.984901 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" podStartSLOduration=69.983307319 podStartE2EDuration="1m9.983307319s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:24.980170749 +0000 UTC m=+96.993449276" watchObservedRunningTime="2025-09-30 16:13:24.983307319 +0000 UTC m=+96.996585836" Sep 30 16:13:24 crc kubenswrapper[4796]: I0930 16:13:24.990220 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" event={"ID":"c239d916-4d86-47a9-b2c5-cac73bcea6bf","Type":"ContainerStarted","Data":"df485d11ff27c1b5da6512546a40301c9b12c906d2f6b9259a7b764992e30e83"} Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.033939 4796 generic.go:334] "Generic (PLEG): container finished" podID="7adf9170-1515-4c07-b471-bcc775bf4327" containerID="542360c40490203007e085bb4278caed4be2080a17772b20541bdb1fce8fd94d" exitCode=0 Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.034298 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" event={"ID":"7adf9170-1515-4c07-b471-bcc775bf4327","Type":"ContainerDied","Data":"542360c40490203007e085bb4278caed4be2080a17772b20541bdb1fce8fd94d"} Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.034330 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" event={"ID":"7adf9170-1515-4c07-b471-bcc775bf4327","Type":"ContainerStarted","Data":"1d9dc6d3d133980c1bd4f454a15c547caf6ad8fa796d833ab0b170b744de82bf"} Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.036148 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.040311 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.041712 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.541696527 +0000 UTC m=+97.554975054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.061883 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" event={"ID":"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7","Type":"ContainerStarted","Data":"ac1a862a79b8ad6e7e40283254434e7fc495dba63a3a39801479ac540093401d"} Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.068062 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-dxc8p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.068114 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dxc8p" podUID="025a66cf-d82d-4590-b9cd-402bc6bc3668" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.074636 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.082284 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.094785 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-sl6k6" podStartSLOduration=6.094764963 podStartE2EDuration="6.094764963s" podCreationTimestamp="2025-09-30 16:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:25.088058341 +0000 UTC m=+97.101336868" watchObservedRunningTime="2025-09-30 16:13:25.094764963 +0000 UTC m=+97.108043490" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.098073 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:25 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:25 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:25 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.098115 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.098447 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-22cb7" podStartSLOduration=70.098437678 podStartE2EDuration="1m10.098437678s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:25.063114939 +0000 UTC m=+97.076393466" watchObservedRunningTime="2025-09-30 16:13:25.098437678 +0000 UTC m=+97.111716205" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.119445 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.141653 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.143524 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.643511725 +0000 UTC m=+97.656790252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.246828 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.249130 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.749100282 +0000 UTC m=+97.762378879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.270178 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" podStartSLOduration=71.270160294 podStartE2EDuration="1m11.270160294s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:25.268005412 +0000 UTC m=+97.281283939" watchObservedRunningTime="2025-09-30 16:13:25.270160294 +0000 UTC m=+97.283438811" Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.353182 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.353531 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.853520115 +0000 UTC m=+97.866798642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.454048 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.454447 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:25.954432278 +0000 UTC m=+97.967710805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.555750 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.556280 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.056264527 +0000 UTC m=+98.069543054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.657322 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.657511 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.157484278 +0000 UTC m=+98.170762805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.657675 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.658054 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.158046504 +0000 UTC m=+98.171325031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.759223 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.760118 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.2600956 +0000 UTC m=+98.273374137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.861793 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.862216 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.362202227 +0000 UTC m=+98.375480754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.963112 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.963646 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.463621624 +0000 UTC m=+98.476900151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:25 crc kubenswrapper[4796]: I0930 16:13:25.963844 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:25 crc kubenswrapper[4796]: E0930 16:13:25.964146 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.464130758 +0000 UTC m=+98.477409285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.068370 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.068790 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.568773518 +0000 UTC m=+98.582052035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.081297 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:26 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:26 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:26 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.081363 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.101309 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" event={"ID":"a540a9ba-ecd7-4490-805b-45ff88d97a63","Type":"ContainerStarted","Data":"e75005ddc6532e6f4c7c08a3396ca1e50a0ec6423855e30533e89067ca5458f6"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.101743 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.107370 4796 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h49fq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.107414 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" podUID="a540a9ba-ecd7-4490-805b-45ff88d97a63" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.124101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" event={"ID":"06101703-e6d8-42e5-8549-7340fbf1a597","Type":"ContainerStarted","Data":"ff2616c791f4c6f4b4e4c2ae79e5a93fd73146e5f1f5bb0de962e5e2435e6822"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.135868 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" podStartSLOduration=71.135847224 podStartE2EDuration="1m11.135847224s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.134463274 +0000 UTC m=+98.147741801" watchObservedRunningTime="2025-09-30 16:13:26.135847224 +0000 UTC m=+98.149125751" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.172234 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.172626 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.672612834 +0000 UTC m=+98.685891371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.181716 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mlhmv" event={"ID":"15579146-527c-4348-9e39-6a9d0ba6533e","Type":"ContainerStarted","Data":"1e4d2a19cf9cf30796b51dc5abbd23ddeb1a15b21a5167082ff6d42b76e8c048"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.187179 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" event={"ID":"3dc25d62-cd1c-45aa-a1c9-740d9e9a5fa7","Type":"ContainerStarted","Data":"69a6608c58ea80dda7acbdfcd7236645ac8d6610c399bdfc8cbdd376fa7d6e15"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.190428 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" event={"ID":"daf84026-ecfa-446d-91dc-e8146d350877","Type":"ContainerStarted","Data":"25f26dbd8cc7dbbbab0543407507160a6d9c0d43779eee5167ebe78e9d9d8288"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.212648 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" podStartSLOduration=72.212630967 podStartE2EDuration="1m12.212630967s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.210683852 +0000 UTC m=+98.223962379" watchObservedRunningTime="2025-09-30 16:13:26.212630967 +0000 UTC m=+98.225909494" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.234487 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" event={"ID":"7881eaa3-77bc-430c-a2a6-d44ada95cc0a","Type":"ContainerStarted","Data":"6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.236035 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.243158 4796 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bbg46 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.243264 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.255321 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" event={"ID":"2528392f-8a0a-4ec7-8360-5c4f99dc14fa","Type":"ContainerStarted","Data":"4466cdef91302f92d7fc3272496856720d0368f03a17525d1e0decebe57ebdd8"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.255378 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" event={"ID":"2528392f-8a0a-4ec7-8360-5c4f99dc14fa","Type":"ContainerStarted","Data":"2b456ce1a39d41514027ed0d312607c8e985df09d3dae686ffce46f3253c9f05"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.272371 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" event={"ID":"c239d916-4d86-47a9-b2c5-cac73bcea6bf","Type":"ContainerStarted","Data":"bb69778d87526dfa9f9fc085fb404148f6aa8fb85203ed5f81a37ff17e1b3e75"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.272431 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" event={"ID":"c239d916-4d86-47a9-b2c5-cac73bcea6bf","Type":"ContainerStarted","Data":"de9ae6ea91a3222c95476df43d785d43ab96fae9cd6fddadb94174a797e7c3d2"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.273430 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.273889 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.275495 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.775469783 +0000 UTC m=+98.788748310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.291505 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pg2pz" podStartSLOduration=71.29148 podStartE2EDuration="1m11.29148s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.288484384 +0000 UTC m=+98.301762911" watchObservedRunningTime="2025-09-30 16:13:26.29148 +0000 UTC m=+98.304758517" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.309165 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" event={"ID":"ecf5a09b-2113-4202-83c1-ad8c77520791","Type":"ContainerStarted","Data":"c471536a485c7ef00f94d01e74ecb9b7d80de1407bacc6bea8f63a04099dcc02"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.309225 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" event={"ID":"ecf5a09b-2113-4202-83c1-ad8c77520791","Type":"ContainerStarted","Data":"29b80ddc09fdb3318788efeec34c3ae9ec8f444958f3900dee2d431e7a599f8e"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.339434 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" event={"ID":"889e3661-2ba6-45ed-a371-3d9b6300db64","Type":"ContainerStarted","Data":"b29056ec4fe6a4c7f5e3796cb0e6db59f91e3c48ac694e22af1baef9460c4959"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.383408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" event={"ID":"49060193-d216-473c-8b67-6eb0d8970dab","Type":"ContainerStarted","Data":"5d0a85b331266bc46961922c77eeea03bbd4984162ae27806b7ec7e3bd01966d"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.383481 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" event={"ID":"49060193-d216-473c-8b67-6eb0d8970dab","Type":"ContainerStarted","Data":"791dd08957a5bc9b5dd04654613e8a3056010cce3b46f16f0fa9ba7b065b0653"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.385529 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbnlg" podStartSLOduration=71.385510186 podStartE2EDuration="1m11.385510186s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.384960331 +0000 UTC m=+98.398238848" watchObservedRunningTime="2025-09-30 16:13:26.385510186 +0000 UTC m=+98.398788713" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.387941 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-s9qpw" podStartSLOduration=71.387931535 podStartE2EDuration="1m11.387931535s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.33978696 +0000 UTC m=+98.353065487" watchObservedRunningTime="2025-09-30 16:13:26.387931535 +0000 UTC m=+98.401210072" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.388262 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.388669 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:26.888650856 +0000 UTC m=+98.901929383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.438469 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" event={"ID":"5496d724-4b33-4818-a317-d1ace3c64aeb","Type":"ContainerStarted","Data":"f4c91f25614ac73555783f7791e415c9823d2dac955114845af940ec9b883da9"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.448992 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" event={"ID":"be7f735b-ccce-4060-915d-58aea627d7f5","Type":"ContainerStarted","Data":"8c338c146db45827acc93987a6f5852b2bcd5b2df1e905be5abee7dcac6edfc1"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.465743 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" event={"ID":"8d34ff06-5475-4d4f-9e69-c0734d22554b","Type":"ContainerStarted","Data":"5d80336db52b5798bea1208c91262818530685f1eb757139a016dfd7347080ea"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.482597 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-j6mpg" event={"ID":"a355257c-502e-495a-97f0-bc65d5f1cfd5","Type":"ContainerStarted","Data":"3562b05397392aa3a9b7cac26abd8f8358b31fc80f4e0f69ac248b850391cc64"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.489130 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.489186 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.497411 4796 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5zv48 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.497486 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" podUID="06101703-e6d8-42e5-8549-7340fbf1a597" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.503685 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.505199 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.005178625 +0000 UTC m=+99.018457152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.529233 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" event={"ID":"cc5983b8-5150-4d8f-8b3a-b45e0bbc10cc","Type":"ContainerStarted","Data":"ba14afd9f866236243869f16351feea4c48a64d5c05d95ca77c30cff626d212c"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.557537 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" event={"ID":"309a1b03-85dc-466d-afdb-8c946fd030e1","Type":"ContainerStarted","Data":"4c1e076c109c670a49ee1d039ffe607dc3655a32b42f7b55094a1f407f94e18d"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.588472 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" event={"ID":"edd79356-997e-4b56-b884-921db4a45d38","Type":"ContainerStarted","Data":"fddd6d7f57cc8b016a6485086d30b510b88150230e3b7d9cca5576867743308d"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.605595 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" podStartSLOduration=72.605576233 podStartE2EDuration="1m12.605576233s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.524050774 +0000 UTC m=+98.537329301" watchObservedRunningTime="2025-09-30 16:13:26.605576233 +0000 UTC m=+98.618854760" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.606737 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.607340 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.107324293 +0000 UTC m=+99.120602820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.615714 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" event={"ID":"f63ad2d2-d528-48e8-99f8-891d91f72ded","Type":"ContainerStarted","Data":"beea58191c5bf60a1beace041b72db5bf8a99f6f710d47c213caba22d19a81c9"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.652695 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" event={"ID":"4c6f0a6e-14ff-45a7-8312-09fd13121728","Type":"ContainerStarted","Data":"f50b6e59b985bd44be3dadf4e529fadda0a0f6de9f6aafe071ada7d096137721"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.689521 4796 generic.go:334] "Generic (PLEG): container finished" podID="47d764ec-58e3-4604-a3b6-259dc3c3fa36" containerID="fc8ba2fd319db3cef92849df912609fd0ce366ba49e4dc356520d320b3b8cf91" exitCode=0 Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.689644 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" event={"ID":"47d764ec-58e3-4604-a3b6-259dc3c3fa36","Type":"ContainerDied","Data":"fc8ba2fd319db3cef92849df912609fd0ce366ba49e4dc356520d320b3b8cf91"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.704911 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" event={"ID":"c9f8bd48-98d2-469a-ad2b-37ad58c7a23d","Type":"ContainerStarted","Data":"0c29a16ee7b200ba6c5e8cedfaed80de8323e1f124541e28398cc679983e1ac7"} Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.707334 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.708226 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.208199045 +0000 UTC m=+99.221477642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.723086 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cps74" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.723468 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" podStartSLOduration=71.72345217 podStartE2EDuration="1m11.72345217s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.722772601 +0000 UTC m=+98.736051128" watchObservedRunningTime="2025-09-30 16:13:26.72345217 +0000 UTC m=+98.736730697" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.724246 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" podStartSLOduration=71.724239583 podStartE2EDuration="1m11.724239583s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.6068647 +0000 UTC m=+98.620143227" watchObservedRunningTime="2025-09-30 16:13:26.724239583 +0000 UTC m=+98.737518110" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.764214 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" podStartSLOduration=71.764197964 podStartE2EDuration="1m11.764197964s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.764169083 +0000 UTC m=+98.777447610" watchObservedRunningTime="2025-09-30 16:13:26.764197964 +0000 UTC m=+98.777476481" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.810823 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.811651 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.311635819 +0000 UTC m=+99.324914346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.847745 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-c9p6v" podStartSLOduration=72.847725281 podStartE2EDuration="1m12.847725281s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.847547645 +0000 UTC m=+98.860826172" watchObservedRunningTime="2025-09-30 16:13:26.847725281 +0000 UTC m=+98.861003808" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.849474 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsrld" podStartSLOduration=71.84946835 podStartE2EDuration="1m11.84946835s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.808603153 +0000 UTC m=+98.821881680" watchObservedRunningTime="2025-09-30 16:13:26.84946835 +0000 UTC m=+98.862746877" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.890146 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m8lpk" podStartSLOduration=71.890124852 podStartE2EDuration="1m11.890124852s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.884643095 +0000 UTC m=+98.897921622" watchObservedRunningTime="2025-09-30 16:13:26.890124852 +0000 UTC m=+98.903403379" Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.912479 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:26 crc kubenswrapper[4796]: E0930 16:13:26.912744 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.412731418 +0000 UTC m=+99.426009945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:26 crc kubenswrapper[4796]: I0930 16:13:26.939953 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cwrnl" podStartSLOduration=71.939922584 podStartE2EDuration="1m11.939922584s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.922216649 +0000 UTC m=+98.935495176" watchObservedRunningTime="2025-09-30 16:13:26.939922584 +0000 UTC m=+98.953201111" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.014608 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.015972 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.515954096 +0000 UTC m=+99.529232623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.039268 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w2nkl" podStartSLOduration=73.039250212 podStartE2EDuration="1m13.039250212s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:26.991774376 +0000 UTC m=+99.005052903" watchObservedRunningTime="2025-09-30 16:13:27.039250212 +0000 UTC m=+99.052528739" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.040181 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mzcjx" podStartSLOduration=72.040175088 podStartE2EDuration="1m12.040175088s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.038337816 +0000 UTC m=+99.051616343" watchObservedRunningTime="2025-09-30 16:13:27.040175088 +0000 UTC m=+99.053453615" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.082061 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:27 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:27 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:27 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.082138 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.083348 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" podStartSLOduration=73.083323621 podStartE2EDuration="1m13.083323621s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.083061973 +0000 UTC m=+99.096340500" watchObservedRunningTime="2025-09-30 16:13:27.083323621 +0000 UTC m=+99.096602148" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.110521 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xm9vj" podStartSLOduration=72.110503287 podStartE2EDuration="1m12.110503287s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.110462876 +0000 UTC m=+99.123741403" watchObservedRunningTime="2025-09-30 16:13:27.110503287 +0000 UTC m=+99.123781814" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.116796 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.117219 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.617199119 +0000 UTC m=+99.630477646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.142330 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-j6mpg" podStartSLOduration=8.142307266 podStartE2EDuration="8.142307266s" podCreationTimestamp="2025-09-30 16:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.138743114 +0000 UTC m=+99.152021641" watchObservedRunningTime="2025-09-30 16:13:27.142307266 +0000 UTC m=+99.155585793" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.219321 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.220108 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.720091718 +0000 UTC m=+99.733370245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.321153 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.321361 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.82133502 +0000 UTC m=+99.834613547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.321565 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.321846 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.821832445 +0000 UTC m=+99.835110972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.422944 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.423157 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.923121348 +0000 UTC m=+99.936399875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.423496 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.423921 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:27.923904 +0000 UTC m=+99.937182527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.477913 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.477964 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.480180 4796 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-gc8lg container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.480250 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" podUID="47d764ec-58e3-4604-a3b6-259dc3c3fa36" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.524782 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.525102 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.025074901 +0000 UTC m=+100.038353428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.626849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.627248 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.127230229 +0000 UTC m=+100.140508756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.711973 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" event={"ID":"edd79356-997e-4b56-b884-921db4a45d38","Type":"ContainerStarted","Data":"3a68d5e2bf35ee2eff5b293e0599171799c7ccc1b72fd622802e2cd7f130c4e4"} Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.714349 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mvv7h" event={"ID":"f63ad2d2-d528-48e8-99f8-891d91f72ded","Type":"ContainerStarted","Data":"437490cdb1dc9a5294515c11b4f645326cc3940d1910028242a4efd390eaae14"} Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.717274 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" event={"ID":"47d764ec-58e3-4604-a3b6-259dc3c3fa36","Type":"ContainerStarted","Data":"d823440791526d9d206b305cff4d34a718639228fd60a2d5869ff6e1b80c4b37"} Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.720593 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" event={"ID":"5496d724-4b33-4818-a317-d1ace3c64aeb","Type":"ContainerStarted","Data":"61c6c68561864c0e3d25c22c905eb7a5dfd6828e18bdc1bc98964dc17445d4aa"} Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.727635 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.727864 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.227831623 +0000 UTC m=+100.241110150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.728035 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.728532 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.228514412 +0000 UTC m=+100.241792939 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.744419 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mlhmv" event={"ID":"15579146-527c-4348-9e39-6a9d0ba6533e","Type":"ContainerStarted","Data":"24c53f4653ac771e3ef8e195014813cad8867e66ef8abb35415eaaffb23f8459"} Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.745109 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.752035 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" event={"ID":"889e3661-2ba6-45ed-a371-3d9b6300db64","Type":"ContainerStarted","Data":"7fbe960fdbd90f1eaa8f27a2b41898af3c230e1dce5be209d19c4fd63d3f4eb4"} Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.753497 4796 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bbg46 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.753550 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.779450 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2r2gh" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.829527 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.832412 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.33238961 +0000 UTC m=+100.345668137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.872838 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-h6hqm" podStartSLOduration=72.872800914 podStartE2EDuration="1m12.872800914s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.798051889 +0000 UTC m=+99.811330416" watchObservedRunningTime="2025-09-30 16:13:27.872800914 +0000 UTC m=+99.886079441" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.909104 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-mlhmv" podStartSLOduration=8.909083951 podStartE2EDuration="8.909083951s" podCreationTimestamp="2025-09-30 16:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.86739517 +0000 UTC m=+99.880673697" watchObservedRunningTime="2025-09-30 16:13:27.909083951 +0000 UTC m=+99.922362478" Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.932890 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:27 crc kubenswrapper[4796]: E0930 16:13:27.933366 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.433345964 +0000 UTC m=+100.446624491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:27 crc kubenswrapper[4796]: I0930 16:13:27.975095 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzrf4" podStartSLOduration=72.975076156 podStartE2EDuration="1m12.975076156s" podCreationTimestamp="2025-09-30 16:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:27.909332818 +0000 UTC m=+99.922611345" watchObservedRunningTime="2025-09-30 16:13:27.975076156 +0000 UTC m=+99.988354683" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.034186 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.034525 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.534473833 +0000 UTC m=+100.547752360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.079448 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:28 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:28 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:28 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.079556 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.082393 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wgjk8"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.083284 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.107222 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgjk8"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.111703 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.135691 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-utilities\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.135783 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-catalog-content\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.135815 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fplfk\" (UniqueName: \"kubernetes.io/projected/c0e5842f-b782-477d-87d7-c677e786c155-kube-api-access-fplfk\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.135861 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.136141 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.636126946 +0000 UTC m=+100.649405473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.194204 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h49fq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.236870 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.237108 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.73707717 +0000 UTC m=+100.750355697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.237179 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.237315 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-utilities\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.237446 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-catalog-content\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.237532 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fplfk\" (UniqueName: \"kubernetes.io/projected/c0e5842f-b782-477d-87d7-c677e786c155-kube-api-access-fplfk\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.237686 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.737665047 +0000 UTC m=+100.750943754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.238309 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-utilities\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.238384 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-catalog-content\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.248802 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pspfv"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.249818 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.252438 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.302659 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fplfk\" (UniqueName: \"kubernetes.io/projected/c0e5842f-b782-477d-87d7-c677e786c155-kube-api-access-fplfk\") pod \"certified-operators-wgjk8\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.330846 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pspfv"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.338585 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.338832 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.838783095 +0000 UTC m=+100.852061632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.339142 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-utilities\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.339319 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfbs6\" (UniqueName: \"kubernetes.io/projected/78afec2a-ad1c-4999-8165-eda2ed82ef5d-kube-api-access-xfbs6\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.339426 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.339594 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-catalog-content\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.339837 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.839809415 +0000 UTC m=+100.853088152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.397708 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.441178 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.441656 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-utilities\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.441712 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfbs6\" (UniqueName: \"kubernetes.io/projected/78afec2a-ad1c-4999-8165-eda2ed82ef5d-kube-api-access-xfbs6\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.441787 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-catalog-content\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.442306 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-catalog-content\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.442484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-utilities\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.442592 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:28.94257139 +0000 UTC m=+100.955849917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.445161 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4mqzq"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.446092 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.527893 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfbs6\" (UniqueName: \"kubernetes.io/projected/78afec2a-ad1c-4999-8165-eda2ed82ef5d-kube-api-access-xfbs6\") pod \"community-operators-pspfv\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.530102 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mqzq"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.542832 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs8sw\" (UniqueName: \"kubernetes.io/projected/0b156b66-9ee3-4b88-9874-acac9e6606a1-kube-api-access-xs8sw\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.542875 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-catalog-content\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.543173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.543231 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-utilities\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.543492 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.043478213 +0000 UTC m=+101.056756730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.572733 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.637750 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2f9rf"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.638673 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.648195 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.648348 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-utilities\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.648384 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs8sw\" (UniqueName: \"kubernetes.io/projected/0b156b66-9ee3-4b88-9874-acac9e6606a1-kube-api-access-xs8sw\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.648410 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-catalog-content\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.649122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-catalog-content\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.649198 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.149182382 +0000 UTC m=+101.162460909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.649415 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-utilities\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.666003 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2f9rf"] Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.706837 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs8sw\" (UniqueName: \"kubernetes.io/projected/0b156b66-9ee3-4b88-9874-acac9e6606a1-kube-api-access-xs8sw\") pod \"certified-operators-4mqzq\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.754957 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.755343 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-catalog-content\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.755385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-utilities\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.755406 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktmhq\" (UniqueName: \"kubernetes.io/projected/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-kube-api-access-ktmhq\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.755710 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.255698945 +0000 UTC m=+101.268977472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.764672 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.839251 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" event={"ID":"889e3661-2ba6-45ed-a371-3d9b6300db64","Type":"ContainerStarted","Data":"236038c424eda8aea976f6f127deccb00d0acaa666cb5775e0e7051228d78684"} Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.847807 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.860034 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.860354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-catalog-content\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.860396 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-utilities\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.860421 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktmhq\" (UniqueName: \"kubernetes.io/projected/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-kube-api-access-ktmhq\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.860713 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.360699535 +0000 UTC m=+101.373978062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.861141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-catalog-content\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.861554 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-utilities\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.947185 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktmhq\" (UniqueName: \"kubernetes.io/projected/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-kube-api-access-ktmhq\") pod \"community-operators-2f9rf\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:28 crc kubenswrapper[4796]: I0930 16:13:28.962455 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:28 crc kubenswrapper[4796]: E0930 16:13:28.962902 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.462887044 +0000 UTC m=+101.476165571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.012636 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.065871 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.067042 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.567014129 +0000 UTC m=+101.580292656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.079836 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:29 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:29 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:29 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.079908 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.163139 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgjk8"] Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.166874 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pspfv"] Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.175963 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.176306 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.676293941 +0000 UTC m=+101.689572468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.275474 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mqzq"] Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.279873 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.288430 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.788404483 +0000 UTC m=+101.801683010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.382741 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.383442 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.883425128 +0000 UTC m=+101.896703655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.384089 4796 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.491707 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.492106 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:29.992091492 +0000 UTC m=+102.005370019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.593464 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.593775 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:30.093762997 +0000 UTC m=+102.107041524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.694666 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.695053 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 16:13:30.19503757 +0000 UTC m=+102.208316097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.728803 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2f9rf"] Sep 30 16:13:29 crc kubenswrapper[4796]: W0930 16:13:29.737762 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaae79f18_b8bd_46d1_8af1_5f2acbaa66ab.slice/crio-25de62b1ae802fc09e8828143514b0267b8f04222cb2292d45d7a673b7f78ede WatchSource:0}: Error finding container 25de62b1ae802fc09e8828143514b0267b8f04222cb2292d45d7a673b7f78ede: Status 404 returned error can't find the container with id 25de62b1ae802fc09e8828143514b0267b8f04222cb2292d45d7a673b7f78ede Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.795694 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:29 crc kubenswrapper[4796]: E0930 16:13:29.796054 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 16:13:30.296032335 +0000 UTC m=+102.309310862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fl5fz" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.849914 4796 generic.go:334] "Generic (PLEG): container finished" podID="c0e5842f-b782-477d-87d7-c677e786c155" containerID="a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd" exitCode=0 Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.850003 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgjk8" event={"ID":"c0e5842f-b782-477d-87d7-c677e786c155","Type":"ContainerDied","Data":"a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.850030 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgjk8" event={"ID":"c0e5842f-b782-477d-87d7-c677e786c155","Type":"ContainerStarted","Data":"7edd1acf1a4ee10ad6f189de9483dc74bcb7be75a84b9b1036ea3bdd4b390414"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.851628 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9rf" event={"ID":"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab","Type":"ContainerStarted","Data":"25de62b1ae802fc09e8828143514b0267b8f04222cb2292d45d7a673b7f78ede"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.851811 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.861161 4796 generic.go:334] "Generic (PLEG): container finished" podID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerID="1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71" exitCode=0 Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.861243 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerDied","Data":"1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.861277 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerStarted","Data":"e8c11238ec3be1f9f8a001af598414cd8514c1596908bf48ff01987350197d27"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.868630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" event={"ID":"889e3661-2ba6-45ed-a371-3d9b6300db64","Type":"ContainerStarted","Data":"3dfcfb98ff67ffaa6149f424edfdc9a13b15560a8a0ceaf2128ed146cfb9f8d1"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.868670 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" event={"ID":"889e3661-2ba6-45ed-a371-3d9b6300db64","Type":"ContainerStarted","Data":"84a48c00825d65046c9690001b9eaacff66a3e05151db4e6116304d4e8c19581"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.872041 4796 generic.go:334] "Generic (PLEG): container finished" podID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerID="65cabb7899a2728ce70424af0ebac1f48a8af14f90f30256b82be183b09d9273" exitCode=0 Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.872117 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mqzq" event={"ID":"0b156b66-9ee3-4b88-9874-acac9e6606a1","Type":"ContainerDied","Data":"65cabb7899a2728ce70424af0ebac1f48a8af14f90f30256b82be183b09d9273"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.872157 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mqzq" event={"ID":"0b156b66-9ee3-4b88-9874-acac9e6606a1","Type":"ContainerStarted","Data":"fb630c1919e851445706f027feaa5bda02a685aaeb918270b9bc7bb45c1f37e9"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.875144 4796 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-30T16:13:29.384107887Z","Handler":null,"Name":""} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.876807 4796 generic.go:334] "Generic (PLEG): container finished" podID="2528392f-8a0a-4ec7-8360-5c4f99dc14fa" containerID="4466cdef91302f92d7fc3272496856720d0368f03a17525d1e0decebe57ebdd8" exitCode=0 Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.877520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" event={"ID":"2528392f-8a0a-4ec7-8360-5c4f99dc14fa","Type":"ContainerDied","Data":"4466cdef91302f92d7fc3272496856720d0368f03a17525d1e0decebe57ebdd8"} Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.882843 4796 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.882881 4796 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.890407 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-fw6t6" podStartSLOduration=10.89038604 podStartE2EDuration="10.89038604s" podCreationTimestamp="2025-09-30 16:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:29.890279337 +0000 UTC m=+101.903557864" watchObservedRunningTime="2025-09-30 16:13:29.89038604 +0000 UTC m=+101.903664587" Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.896675 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.900481 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 16:13:29 crc kubenswrapper[4796]: I0930 16:13:29.998062 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.010387 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.010432 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.028854 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lhk59"] Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.030049 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.033715 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.048026 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fl5fz\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.077527 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhk59"] Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.081064 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:30 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:30 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:30 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.081138 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.122841 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.200334 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-utilities\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.200442 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-catalog-content\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.200488 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skfm9\" (UniqueName: \"kubernetes.io/projected/25f6d193-b5f7-49cc-acf0-1698e5a15de7-kube-api-access-skfm9\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.301858 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-catalog-content\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.302373 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skfm9\" (UniqueName: \"kubernetes.io/projected/25f6d193-b5f7-49cc-acf0-1698e5a15de7-kube-api-access-skfm9\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.302436 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-utilities\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.303169 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-catalog-content\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.303262 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-utilities\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.327808 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skfm9\" (UniqueName: \"kubernetes.io/projected/25f6d193-b5f7-49cc-acf0-1698e5a15de7-kube-api-access-skfm9\") pod \"redhat-marketplace-lhk59\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.345610 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.435790 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hldvs"] Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.439514 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.457570 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldvs"] Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.507077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-catalog-content\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.507180 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-utilities\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.507231 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6s8r\" (UniqueName: \"kubernetes.io/projected/ccbaff9e-7239-4b52-9123-fda87849ccd4-kube-api-access-l6s8r\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.545353 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhk59"] Sep 30 16:13:30 crc kubenswrapper[4796]: W0930 16:13:30.553105 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25f6d193_b5f7_49cc_acf0_1698e5a15de7.slice/crio-584323153f468cb319e1939c4420075d6e0f7e8c4bdfd3ebdc073c0c92c7e719 WatchSource:0}: Error finding container 584323153f468cb319e1939c4420075d6e0f7e8c4bdfd3ebdc073c0c92c7e719: Status 404 returned error can't find the container with id 584323153f468cb319e1939c4420075d6e0f7e8c4bdfd3ebdc073c0c92c7e719 Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.576383 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fl5fz"] Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.608781 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-catalog-content\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.609287 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-utilities\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.609333 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6s8r\" (UniqueName: \"kubernetes.io/projected/ccbaff9e-7239-4b52-9123-fda87849ccd4-kube-api-access-l6s8r\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.609451 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-catalog-content\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.609711 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-utilities\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.629805 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6s8r\" (UniqueName: \"kubernetes.io/projected/ccbaff9e-7239-4b52-9123-fda87849ccd4-kube-api-access-l6s8r\") pod \"redhat-marketplace-hldvs\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.745719 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.763479 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.890299 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" event={"ID":"8aba7c6e-3d24-4217-9687-93a535675dbc","Type":"ContainerStarted","Data":"c2da43cb5e55daaa18fabf2be3caf53ed82144610d785157935b601e098ebe4b"} Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.890355 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" event={"ID":"8aba7c6e-3d24-4217-9687-93a535675dbc","Type":"ContainerStarted","Data":"123e66d98bbb2b844640ae619e030b4ead9bf588bdaca5c91cc93601692cad91"} Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.890448 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.904366 4796 generic.go:334] "Generic (PLEG): container finished" podID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerID="7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21" exitCode=0 Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.904443 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhk59" event={"ID":"25f6d193-b5f7-49cc-acf0-1698e5a15de7","Type":"ContainerDied","Data":"7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21"} Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.904470 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhk59" event={"ID":"25f6d193-b5f7-49cc-acf0-1698e5a15de7","Type":"ContainerStarted","Data":"584323153f468cb319e1939c4420075d6e0f7e8c4bdfd3ebdc073c0c92c7e719"} Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.916366 4796 generic.go:334] "Generic (PLEG): container finished" podID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerID="0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c" exitCode=0 Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.916483 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9rf" event={"ID":"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab","Type":"ContainerDied","Data":"0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c"} Sep 30 16:13:30 crc kubenswrapper[4796]: I0930 16:13:30.929508 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" podStartSLOduration=76.929483125 podStartE2EDuration="1m16.929483125s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:30.909139954 +0000 UTC m=+102.922418481" watchObservedRunningTime="2025-09-30 16:13:30.929483125 +0000 UTC m=+102.942761652" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.031202 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.032397 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.034135 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.034307 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.040463 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.078464 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:31 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:31 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:31 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.078538 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.125581 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldvs"] Sep 30 16:13:31 crc kubenswrapper[4796]: W0930 16:13:31.134893 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccbaff9e_7239_4b52_9123_fda87849ccd4.slice/crio-ab0da10d6343368372d53e490558dff68f2c91e5a0ccf67b356288644748b385 WatchSource:0}: Error finding container ab0da10d6343368372d53e490558dff68f2c91e5a0ccf67b356288644748b385: Status 404 returned error can't find the container with id ab0da10d6343368372d53e490558dff68f2c91e5a0ccf67b356288644748b385 Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.203101 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.226720 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-secret-volume\") pod \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.226811 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-config-volume\") pod \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.226847 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5j6m\" (UniqueName: \"kubernetes.io/projected/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-kube-api-access-l5j6m\") pod \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\" (UID: \"2528392f-8a0a-4ec7-8360-5c4f99dc14fa\") " Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.226997 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054d1bfa-f102-45cc-8e25-36f77646a73d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.227037 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/054d1bfa-f102-45cc-8e25-36f77646a73d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.228595 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-config-volume" (OuterVolumeSpecName: "config-volume") pod "2528392f-8a0a-4ec7-8360-5c4f99dc14fa" (UID: "2528392f-8a0a-4ec7-8360-5c4f99dc14fa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.252675 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-kube-api-access-l5j6m" (OuterVolumeSpecName: "kube-api-access-l5j6m") pod "2528392f-8a0a-4ec7-8360-5c4f99dc14fa" (UID: "2528392f-8a0a-4ec7-8360-5c4f99dc14fa"). InnerVolumeSpecName "kube-api-access-l5j6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.260697 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2528392f-8a0a-4ec7-8360-5c4f99dc14fa" (UID: "2528392f-8a0a-4ec7-8360-5c4f99dc14fa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.328128 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054d1bfa-f102-45cc-8e25-36f77646a73d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.328189 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/054d1bfa-f102-45cc-8e25-36f77646a73d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.328263 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5j6m\" (UniqueName: \"kubernetes.io/projected/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-kube-api-access-l5j6m\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.328275 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.328284 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2528392f-8a0a-4ec7-8360-5c4f99dc14fa-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.328324 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/054d1bfa-f102-45cc-8e25-36f77646a73d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.347656 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054d1bfa-f102-45cc-8e25-36f77646a73d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.362938 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.439019 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8xs5d"] Sep 30 16:13:31 crc kubenswrapper[4796]: E0930 16:13:31.439750 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2528392f-8a0a-4ec7-8360-5c4f99dc14fa" containerName="collect-profiles" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.439772 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2528392f-8a0a-4ec7-8360-5c4f99dc14fa" containerName="collect-profiles" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.439907 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2528392f-8a0a-4ec7-8360-5c4f99dc14fa" containerName="collect-profiles" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.440869 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.444842 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.447169 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xs5d"] Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.494752 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.501361 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5zv48" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.533054 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-utilities\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.533124 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-catalog-content\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.533212 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st2ww\" (UniqueName: \"kubernetes.io/projected/a77743d1-a6ef-4741-83d2-a102d2a256dc-kube-api-access-st2ww\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.634294 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-utilities\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.634369 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-catalog-content\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.634431 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st2ww\" (UniqueName: \"kubernetes.io/projected/a77743d1-a6ef-4741-83d2-a102d2a256dc-kube-api-access-st2ww\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.634880 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-utilities\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.635311 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-catalog-content\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.656574 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st2ww\" (UniqueName: \"kubernetes.io/projected/a77743d1-a6ef-4741-83d2-a102d2a256dc-kube-api-access-st2ww\") pod \"redhat-operators-8xs5d\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.766760 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.831104 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.834885 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-25vk7"] Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.838879 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:31 crc kubenswrapper[4796]: W0930 16:13:31.846294 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod054d1bfa_f102_45cc_8e25_36f77646a73d.slice/crio-9a8b5d4b001d08daa7b34d926f24a5dfcc383df5bcaac2a98c320f59385cf019 WatchSource:0}: Error finding container 9a8b5d4b001d08daa7b34d926f24a5dfcc383df5bcaac2a98c320f59385cf019: Status 404 returned error can't find the container with id 9a8b5d4b001d08daa7b34d926f24a5dfcc383df5bcaac2a98c320f59385cf019 Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.852215 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25vk7"] Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934029 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934060 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z" event={"ID":"2528392f-8a0a-4ec7-8360-5c4f99dc14fa","Type":"ContainerDied","Data":"2b456ce1a39d41514027ed0d312607c8e985df09d3dae686ffce46f3253c9f05"} Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934114 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b456ce1a39d41514027ed0d312607c8e985df09d3dae686ffce46f3253c9f05" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934224 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-dxc8p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934257 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dxc8p" podUID="025a66cf-d82d-4590-b9cd-402bc6bc3668" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934410 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-dxc8p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.934439 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dxc8p" podUID="025a66cf-d82d-4590-b9cd-402bc6bc3668" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.942758 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g6ft\" (UniqueName: \"kubernetes.io/projected/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-kube-api-access-4g6ft\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.942819 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-utilities\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.942858 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-catalog-content\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.951747 4796 generic.go:334] "Generic (PLEG): container finished" podID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerID="7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04" exitCode=0 Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.951972 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldvs" event={"ID":"ccbaff9e-7239-4b52-9123-fda87849ccd4","Type":"ContainerDied","Data":"7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04"} Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.952028 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldvs" event={"ID":"ccbaff9e-7239-4b52-9123-fda87849ccd4","Type":"ContainerStarted","Data":"ab0da10d6343368372d53e490558dff68f2c91e5a0ccf67b356288644748b385"} Sep 30 16:13:31 crc kubenswrapper[4796]: I0930 16:13:31.966639 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"054d1bfa-f102-45cc-8e25-36f77646a73d","Type":"ContainerStarted","Data":"9a8b5d4b001d08daa7b34d926f24a5dfcc383df5bcaac2a98c320f59385cf019"} Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.033529 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.033955 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.034674 4796 patch_prober.go:28] interesting pod/console-f9d7485db-cch5s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.034752 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-cch5s" podUID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.047615 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g6ft\" (UniqueName: \"kubernetes.io/projected/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-kube-api-access-4g6ft\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.047678 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-utilities\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.047744 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-catalog-content\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.048632 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-catalog-content\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.049141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-utilities\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.069364 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g6ft\" (UniqueName: \"kubernetes.io/projected/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-kube-api-access-4g6ft\") pod \"redhat-operators-25vk7\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.075768 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.080567 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:32 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:32 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:32 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.080633 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.122100 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xs5d"] Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.191936 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.489396 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25vk7"] Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.496056 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.505284 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gc8lg" Sep 30 16:13:32 crc kubenswrapper[4796]: I0930 16:13:32.993720 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"054d1bfa-f102-45cc-8e25-36f77646a73d","Type":"ContainerStarted","Data":"ffde3b669acafca97325f9d2a4b88efb8a30d856835036e230fb8f84e9875817"} Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.009185 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerID="b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc" exitCode=0 Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.009263 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25vk7" event={"ID":"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075","Type":"ContainerDied","Data":"b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc"} Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.009291 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25vk7" event={"ID":"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075","Type":"ContainerStarted","Data":"f20fcbf64403d082cf4b930044c56af45dfaa819fe715b7762e7d5defa2a1aeb"} Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.015912 4796 generic.go:334] "Generic (PLEG): container finished" podID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerID="4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26" exitCode=0 Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.016901 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerDied","Data":"4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26"} Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.016926 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerStarted","Data":"4851808ac3ef5e9980c59344d82c4ae128766562cb9a4d8f445fd7ffce135962"} Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.087108 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:33 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:33 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:33 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.087159 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.185378 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.207606 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1683cfa-0a4a-4f40-bbbf-6675107ecbf9-metrics-certs\") pod \"network-metrics-daemon-tl9z9\" (UID: \"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9\") " pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.360578 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tl9z9" Sep 30 16:13:33 crc kubenswrapper[4796]: I0930 16:13:33.749595 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tl9z9"] Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.052351 4796 generic.go:334] "Generic (PLEG): container finished" podID="054d1bfa-f102-45cc-8e25-36f77646a73d" containerID="ffde3b669acafca97325f9d2a4b88efb8a30d856835036e230fb8f84e9875817" exitCode=0 Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.052447 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"054d1bfa-f102-45cc-8e25-36f77646a73d","Type":"ContainerDied","Data":"ffde3b669acafca97325f9d2a4b88efb8a30d856835036e230fb8f84e9875817"} Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.067243 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" event={"ID":"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9","Type":"ContainerStarted","Data":"c5a6c6b1402ab0d34740e69e1d88abc7258849ff57c2d18da1949422f07fa9fb"} Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.081083 4796 patch_prober.go:28] interesting pod/router-default-5444994796-6r92c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 16:13:34 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Sep 30 16:13:34 crc kubenswrapper[4796]: [+]process-running ok Sep 30 16:13:34 crc kubenswrapper[4796]: healthz check failed Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.082356 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6r92c" podUID="fb586c4b-79c4-4747-b9d4-8ff7e21e03f2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.463147 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.518150 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054d1bfa-f102-45cc-8e25-36f77646a73d-kube-api-access\") pod \"054d1bfa-f102-45cc-8e25-36f77646a73d\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.518590 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/054d1bfa-f102-45cc-8e25-36f77646a73d-kubelet-dir\") pod \"054d1bfa-f102-45cc-8e25-36f77646a73d\" (UID: \"054d1bfa-f102-45cc-8e25-36f77646a73d\") " Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.518988 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/054d1bfa-f102-45cc-8e25-36f77646a73d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "054d1bfa-f102-45cc-8e25-36f77646a73d" (UID: "054d1bfa-f102-45cc-8e25-36f77646a73d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.551276 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/054d1bfa-f102-45cc-8e25-36f77646a73d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "054d1bfa-f102-45cc-8e25-36f77646a73d" (UID: "054d1bfa-f102-45cc-8e25-36f77646a73d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.560860 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-mlhmv" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.626674 4796 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/054d1bfa-f102-45cc-8e25-36f77646a73d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.626703 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054d1bfa-f102-45cc-8e25-36f77646a73d-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:34 crc kubenswrapper[4796]: I0930 16:13:34.865830 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.084531 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"054d1bfa-f102-45cc-8e25-36f77646a73d","Type":"ContainerDied","Data":"9a8b5d4b001d08daa7b34d926f24a5dfcc383df5bcaac2a98c320f59385cf019"} Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.084567 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a8b5d4b001d08daa7b34d926f24a5dfcc383df5bcaac2a98c320f59385cf019" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.084634 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.090225 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.095943 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" event={"ID":"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9","Type":"ContainerStarted","Data":"80faa19af65ac4eb1dad050620c4ec4adf660ad3032103e9b2b0175fec581891"} Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.098476 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-6r92c" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.150710 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 16:13:35 crc kubenswrapper[4796]: E0930 16:13:35.150935 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054d1bfa-f102-45cc-8e25-36f77646a73d" containerName="pruner" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.150947 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="054d1bfa-f102-45cc-8e25-36f77646a73d" containerName="pruner" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.151063 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="054d1bfa-f102-45cc-8e25-36f77646a73d" containerName="pruner" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.151448 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.154396 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.154582 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.169055 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.240256 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f05657c8-ad27-4b34-96e6-c41462e82358-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.240341 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f05657c8-ad27-4b34-96e6-c41462e82358-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.341609 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f05657c8-ad27-4b34-96e6-c41462e82358-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.341749 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f05657c8-ad27-4b34-96e6-c41462e82358-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.341905 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f05657c8-ad27-4b34-96e6-c41462e82358-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.360586 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f05657c8-ad27-4b34-96e6-c41462e82358-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:35 crc kubenswrapper[4796]: I0930 16:13:35.492774 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:36 crc kubenswrapper[4796]: I0930 16:13:36.029782 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 16:13:36 crc kubenswrapper[4796]: W0930 16:13:36.060726 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf05657c8_ad27_4b34_96e6_c41462e82358.slice/crio-b1e6b41193cffd0052d069c0bb6fe9ed3306a714a6e7aa37d49d2d94dbb5b180 WatchSource:0}: Error finding container b1e6b41193cffd0052d069c0bb6fe9ed3306a714a6e7aa37d49d2d94dbb5b180: Status 404 returned error can't find the container with id b1e6b41193cffd0052d069c0bb6fe9ed3306a714a6e7aa37d49d2d94dbb5b180 Sep 30 16:13:36 crc kubenswrapper[4796]: I0930 16:13:36.193556 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f05657c8-ad27-4b34-96e6-c41462e82358","Type":"ContainerStarted","Data":"b1e6b41193cffd0052d069c0bb6fe9ed3306a714a6e7aa37d49d2d94dbb5b180"} Sep 30 16:13:36 crc kubenswrapper[4796]: I0930 16:13:36.226177 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tl9z9" event={"ID":"d1683cfa-0a4a-4f40-bbbf-6675107ecbf9","Type":"ContainerStarted","Data":"72a1e4a28a8af16afc031f72e9fed1498a196fdc48e98ac9348da1e70b905043"} Sep 30 16:13:36 crc kubenswrapper[4796]: I0930 16:13:36.241422 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-tl9z9" podStartSLOduration=82.241374999 podStartE2EDuration="1m22.241374999s" podCreationTimestamp="2025-09-30 16:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:36.239730532 +0000 UTC m=+108.253009059" watchObservedRunningTime="2025-09-30 16:13:36.241374999 +0000 UTC m=+108.254653516" Sep 30 16:13:37 crc kubenswrapper[4796]: I0930 16:13:37.259347 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f05657c8-ad27-4b34-96e6-c41462e82358","Type":"ContainerStarted","Data":"cf6a2351e9caa37e6ce64f408c5be855965824a45fb43358a98ae7f7070f61e2"} Sep 30 16:13:37 crc kubenswrapper[4796]: I0930 16:13:37.281489 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.281459661 podStartE2EDuration="2.281459661s" podCreationTimestamp="2025-09-30 16:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:13:37.279610358 +0000 UTC m=+109.292888885" watchObservedRunningTime="2025-09-30 16:13:37.281459661 +0000 UTC m=+109.294738188" Sep 30 16:13:38 crc kubenswrapper[4796]: I0930 16:13:38.276276 4796 generic.go:334] "Generic (PLEG): container finished" podID="f05657c8-ad27-4b34-96e6-c41462e82358" containerID="cf6a2351e9caa37e6ce64f408c5be855965824a45fb43358a98ae7f7070f61e2" exitCode=0 Sep 30 16:13:38 crc kubenswrapper[4796]: I0930 16:13:38.276366 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f05657c8-ad27-4b34-96e6-c41462e82358","Type":"ContainerDied","Data":"cf6a2351e9caa37e6ce64f408c5be855965824a45fb43358a98ae7f7070f61e2"} Sep 30 16:13:41 crc kubenswrapper[4796]: I0930 16:13:41.941048 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dxc8p" Sep 30 16:13:42 crc kubenswrapper[4796]: I0930 16:13:42.022232 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:42 crc kubenswrapper[4796]: I0930 16:13:42.026843 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.256516 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.330382 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f05657c8-ad27-4b34-96e6-c41462e82358","Type":"ContainerDied","Data":"b1e6b41193cffd0052d069c0bb6fe9ed3306a714a6e7aa37d49d2d94dbb5b180"} Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.330789 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1e6b41193cffd0052d069c0bb6fe9ed3306a714a6e7aa37d49d2d94dbb5b180" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.330645 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.420201 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f05657c8-ad27-4b34-96e6-c41462e82358-kubelet-dir\") pod \"f05657c8-ad27-4b34-96e6-c41462e82358\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.420280 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f05657c8-ad27-4b34-96e6-c41462e82358-kube-api-access\") pod \"f05657c8-ad27-4b34-96e6-c41462e82358\" (UID: \"f05657c8-ad27-4b34-96e6-c41462e82358\") " Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.420442 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f05657c8-ad27-4b34-96e6-c41462e82358-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f05657c8-ad27-4b34-96e6-c41462e82358" (UID: "f05657c8-ad27-4b34-96e6-c41462e82358"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.420777 4796 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f05657c8-ad27-4b34-96e6-c41462e82358-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.439147 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05657c8-ad27-4b34-96e6-c41462e82358-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f05657c8-ad27-4b34-96e6-c41462e82358" (UID: "f05657c8-ad27-4b34-96e6-c41462e82358"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:13:44 crc kubenswrapper[4796]: I0930 16:13:44.522043 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f05657c8-ad27-4b34-96e6-c41462e82358-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 16:13:50 crc kubenswrapper[4796]: I0930 16:13:50.129796 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:14:01 crc kubenswrapper[4796]: E0930 16:14:01.473912 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 16:14:01 crc kubenswrapper[4796]: E0930 16:14:01.475104 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-skfm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lhk59_openshift-marketplace(25f6d193-b5f7-49cc-acf0-1698e5a15de7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:14:01 crc kubenswrapper[4796]: E0930 16:14:01.477278 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lhk59" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" Sep 30 16:14:02 crc kubenswrapper[4796]: I0930 16:14:02.834958 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6xlsv" Sep 30 16:14:03 crc kubenswrapper[4796]: E0930 16:14:03.657298 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lhk59" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" Sep 30 16:14:03 crc kubenswrapper[4796]: E0930 16:14:03.750469 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 16:14:03 crc kubenswrapper[4796]: E0930 16:14:03.751085 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6s8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hldvs_openshift-marketplace(ccbaff9e-7239-4b52-9123-fda87849ccd4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:14:03 crc kubenswrapper[4796]: E0930 16:14:03.752264 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hldvs" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.759324 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hldvs" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.838802 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.839018 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xfbs6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pspfv_openshift-marketplace(78afec2a-ad1c-4999-8165-eda2ed82ef5d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.840267 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pspfv" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.862757 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.862897 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xs8sw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4mqzq_openshift-marketplace(0b156b66-9ee3-4b88-9874-acac9e6606a1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.864001 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4mqzq" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.884335 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.884528 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ktmhq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2f9rf_openshift-marketplace(aae79f18-b8bd-46d1-8af1-5f2acbaa66ab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.885759 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2f9rf" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.910785 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.911339 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4g6ft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-25vk7_openshift-marketplace(8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:14:04 crc kubenswrapper[4796]: E0930 16:14:04.913042 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-25vk7" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" Sep 30 16:14:05 crc kubenswrapper[4796]: I0930 16:14:05.456931 4796 generic.go:334] "Generic (PLEG): container finished" podID="c0e5842f-b782-477d-87d7-c677e786c155" containerID="4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b" exitCode=0 Sep 30 16:14:05 crc kubenswrapper[4796]: I0930 16:14:05.457110 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgjk8" event={"ID":"c0e5842f-b782-477d-87d7-c677e786c155","Type":"ContainerDied","Data":"4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b"} Sep 30 16:14:05 crc kubenswrapper[4796]: I0930 16:14:05.462510 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerStarted","Data":"a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185"} Sep 30 16:14:05 crc kubenswrapper[4796]: E0930 16:14:05.464813 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pspfv" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" Sep 30 16:14:05 crc kubenswrapper[4796]: E0930 16:14:05.464922 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4mqzq" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" Sep 30 16:14:05 crc kubenswrapper[4796]: E0930 16:14:05.465331 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2f9rf" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" Sep 30 16:14:05 crc kubenswrapper[4796]: E0930 16:14:05.466237 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-25vk7" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" Sep 30 16:14:06 crc kubenswrapper[4796]: I0930 16:14:06.466826 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgjk8" event={"ID":"c0e5842f-b782-477d-87d7-c677e786c155","Type":"ContainerStarted","Data":"5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f"} Sep 30 16:14:06 crc kubenswrapper[4796]: I0930 16:14:06.470237 4796 generic.go:334] "Generic (PLEG): container finished" podID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerID="a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185" exitCode=0 Sep 30 16:14:06 crc kubenswrapper[4796]: I0930 16:14:06.470292 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerDied","Data":"a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185"} Sep 30 16:14:06 crc kubenswrapper[4796]: I0930 16:14:06.495872 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wgjk8" podStartSLOduration=2.369760505 podStartE2EDuration="38.495846902s" podCreationTimestamp="2025-09-30 16:13:28 +0000 UTC" firstStartedPulling="2025-09-30 16:13:29.851458728 +0000 UTC m=+101.864737265" lastFinishedPulling="2025-09-30 16:14:05.977545135 +0000 UTC m=+137.990823662" observedRunningTime="2025-09-30 16:14:06.492360922 +0000 UTC m=+138.505639519" watchObservedRunningTime="2025-09-30 16:14:06.495846902 +0000 UTC m=+138.509125429" Sep 30 16:14:07 crc kubenswrapper[4796]: I0930 16:14:07.478101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerStarted","Data":"762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565"} Sep 30 16:14:07 crc kubenswrapper[4796]: I0930 16:14:07.502582 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8xs5d" podStartSLOduration=2.652736039 podStartE2EDuration="36.50256278s" podCreationTimestamp="2025-09-30 16:13:31 +0000 UTC" firstStartedPulling="2025-09-30 16:13:33.02266145 +0000 UTC m=+105.035939977" lastFinishedPulling="2025-09-30 16:14:06.872488191 +0000 UTC m=+138.885766718" observedRunningTime="2025-09-30 16:14:07.502253581 +0000 UTC m=+139.515532108" watchObservedRunningTime="2025-09-30 16:14:07.50256278 +0000 UTC m=+139.515841307" Sep 30 16:14:08 crc kubenswrapper[4796]: I0930 16:14:08.398247 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:14:08 crc kubenswrapper[4796]: I0930 16:14:08.398298 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:14:08 crc kubenswrapper[4796]: I0930 16:14:08.541600 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:14:11 crc kubenswrapper[4796]: I0930 16:14:11.767486 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:14:11 crc kubenswrapper[4796]: I0930 16:14:11.768073 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:14:12 crc kubenswrapper[4796]: I0930 16:14:12.834801 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8xs5d" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="registry-server" probeResult="failure" output=< Sep 30 16:14:12 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Sep 30 16:14:12 crc kubenswrapper[4796]: > Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.572253 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.573095 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.575674 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.576208 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.593304 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.674081 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.674169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.675925 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.686341 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.699191 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.701857 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.703757 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.713198 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.728734 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:14:15 crc kubenswrapper[4796]: I0930 16:14:15.991710 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 16:14:18 crc kubenswrapper[4796]: W0930 16:14:18.298044 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-59ad94fe9ec0d64736f4eb9ea072feab42cfed7a478db8a8ff73793b56526a6e WatchSource:0}: Error finding container 59ad94fe9ec0d64736f4eb9ea072feab42cfed7a478db8a8ff73793b56526a6e: Status 404 returned error can't find the container with id 59ad94fe9ec0d64736f4eb9ea072feab42cfed7a478db8a8ff73793b56526a6e Sep 30 16:14:18 crc kubenswrapper[4796]: I0930 16:14:18.478865 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:14:18 crc kubenswrapper[4796]: I0930 16:14:18.549603 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"125b3ef044bbad7bd1e152417fb401a42f0809d235663613e75c7b32cbc2a7c6"} Sep 30 16:14:18 crc kubenswrapper[4796]: I0930 16:14:18.549666 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ed8673eea0281ed03fe6e04c73f0d39e679266a083664e963a0d96e4f029e539"} Sep 30 16:14:18 crc kubenswrapper[4796]: I0930 16:14:18.550775 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"59ad94fe9ec0d64736f4eb9ea072feab42cfed7a478db8a8ff73793b56526a6e"} Sep 30 16:14:18 crc kubenswrapper[4796]: I0930 16:14:18.552110 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6ae762f5d5630cef3f1e4096a0e6e2fbe23ae3a37f4b4e0accbfe5d7456e0ef1"} Sep 30 16:14:18 crc kubenswrapper[4796]: I0930 16:14:18.552142 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"44064dcd6c87e2dc049fb4a26724535a9266d3d27bb90627018e64bf5a8e7ccd"} Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.563668 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerID="be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc" exitCode=0 Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.563758 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25vk7" event={"ID":"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075","Type":"ContainerDied","Data":"be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc"} Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.567167 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"14f25985a960642016b01c7c3605dc3e3ca13fecf77ef1d61f435e4daf14fbfc"} Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.569206 4796 generic.go:334] "Generic (PLEG): container finished" podID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerID="db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438" exitCode=0 Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.569270 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldvs" event={"ID":"ccbaff9e-7239-4b52-9123-fda87849ccd4","Type":"ContainerDied","Data":"db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438"} Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.572064 4796 generic.go:334] "Generic (PLEG): container finished" podID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerID="35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e" exitCode=0 Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.572124 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhk59" event={"ID":"25f6d193-b5f7-49cc-acf0-1698e5a15de7","Type":"ContainerDied","Data":"35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e"} Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.574225 4796 generic.go:334] "Generic (PLEG): container finished" podID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerID="84dc4a0c1d54ee5ab9c4960779d72e38a6fd4a82f44786d24ce2fee939475660" exitCode=0 Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.574252 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mqzq" event={"ID":"0b156b66-9ee3-4b88-9874-acac9e6606a1","Type":"ContainerDied","Data":"84dc4a0c1d54ee5ab9c4960779d72e38a6fd4a82f44786d24ce2fee939475660"} Sep 30 16:14:19 crc kubenswrapper[4796]: I0930 16:14:19.574611 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:14:20 crc kubenswrapper[4796]: E0930 16:14:20.105725 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaae79f18_b8bd_46d1_8af1_5f2acbaa66ab.slice/crio-conmon-7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.582861 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mqzq" event={"ID":"0b156b66-9ee3-4b88-9874-acac9e6606a1","Type":"ContainerStarted","Data":"e4c8cefed2941e393b39c68bf851c80441cd433dcba318e92b4464735f94e732"} Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.585167 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25vk7" event={"ID":"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075","Type":"ContainerStarted","Data":"ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0"} Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.587103 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldvs" event={"ID":"ccbaff9e-7239-4b52-9123-fda87849ccd4","Type":"ContainerStarted","Data":"2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e"} Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.588676 4796 generic.go:334] "Generic (PLEG): container finished" podID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerID="7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3" exitCode=0 Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.588733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9rf" event={"ID":"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab","Type":"ContainerDied","Data":"7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3"} Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.592041 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerStarted","Data":"9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17"} Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.595330 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhk59" event={"ID":"25f6d193-b5f7-49cc-acf0-1698e5a15de7","Type":"ContainerStarted","Data":"008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060"} Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.637735 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lhk59" podStartSLOduration=1.427304026 podStartE2EDuration="50.637715344s" podCreationTimestamp="2025-09-30 16:13:30 +0000 UTC" firstStartedPulling="2025-09-30 16:13:30.908252658 +0000 UTC m=+102.921531195" lastFinishedPulling="2025-09-30 16:14:20.118663986 +0000 UTC m=+152.131942513" observedRunningTime="2025-09-30 16:14:20.636287783 +0000 UTC m=+152.649566320" watchObservedRunningTime="2025-09-30 16:14:20.637715344 +0000 UTC m=+152.650993871" Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.639011 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-25vk7" podStartSLOduration=2.459273174 podStartE2EDuration="49.639006771s" podCreationTimestamp="2025-09-30 16:13:31 +0000 UTC" firstStartedPulling="2025-09-30 16:13:33.014421375 +0000 UTC m=+105.027699902" lastFinishedPulling="2025-09-30 16:14:20.194154972 +0000 UTC m=+152.207433499" observedRunningTime="2025-09-30 16:14:20.607510311 +0000 UTC m=+152.620788838" watchObservedRunningTime="2025-09-30 16:14:20.639006771 +0000 UTC m=+152.652285298" Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.659969 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hldvs" podStartSLOduration=2.441599741 podStartE2EDuration="50.659950439s" podCreationTimestamp="2025-09-30 16:13:30 +0000 UTC" firstStartedPulling="2025-09-30 16:13:31.954695581 +0000 UTC m=+103.967974098" lastFinishedPulling="2025-09-30 16:14:20.173046269 +0000 UTC m=+152.186324796" observedRunningTime="2025-09-30 16:14:20.656106729 +0000 UTC m=+152.669385256" watchObservedRunningTime="2025-09-30 16:14:20.659950439 +0000 UTC m=+152.673228966" Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.763668 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:14:20 crc kubenswrapper[4796]: I0930 16:14:20.763723 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:14:21 crc kubenswrapper[4796]: I0930 16:14:21.601635 4796 generic.go:334] "Generic (PLEG): container finished" podID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerID="9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17" exitCode=0 Sep 30 16:14:21 crc kubenswrapper[4796]: I0930 16:14:21.601710 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerDied","Data":"9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17"} Sep 30 16:14:21 crc kubenswrapper[4796]: I0930 16:14:21.622303 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4mqzq" podStartSLOduration=3.059317123 podStartE2EDuration="53.62228579s" podCreationTimestamp="2025-09-30 16:13:28 +0000 UTC" firstStartedPulling="2025-09-30 16:13:29.873785146 +0000 UTC m=+101.887063693" lastFinishedPulling="2025-09-30 16:14:20.436753833 +0000 UTC m=+152.450032360" observedRunningTime="2025-09-30 16:14:21.621659622 +0000 UTC m=+153.634938149" watchObservedRunningTime="2025-09-30 16:14:21.62228579 +0000 UTC m=+153.635564317" Sep 30 16:14:21 crc kubenswrapper[4796]: I0930 16:14:21.804903 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-hldvs" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="registry-server" probeResult="failure" output=< Sep 30 16:14:21 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Sep 30 16:14:21 crc kubenswrapper[4796]: > Sep 30 16:14:21 crc kubenswrapper[4796]: I0930 16:14:21.815490 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:14:21 crc kubenswrapper[4796]: I0930 16:14:21.855826 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:14:22 crc kubenswrapper[4796]: I0930 16:14:22.192115 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:14:22 crc kubenswrapper[4796]: I0930 16:14:22.192651 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:14:22 crc kubenswrapper[4796]: I0930 16:14:22.609504 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9rf" event={"ID":"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab","Type":"ContainerStarted","Data":"c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb"} Sep 30 16:14:22 crc kubenswrapper[4796]: I0930 16:14:22.612800 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerStarted","Data":"a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486"} Sep 30 16:14:22 crc kubenswrapper[4796]: I0930 16:14:22.655418 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2f9rf" podStartSLOduration=3.652824789 podStartE2EDuration="54.655394493s" podCreationTimestamp="2025-09-30 16:13:28 +0000 UTC" firstStartedPulling="2025-09-30 16:13:30.923117813 +0000 UTC m=+102.936396340" lastFinishedPulling="2025-09-30 16:14:21.925687527 +0000 UTC m=+153.938966044" observedRunningTime="2025-09-30 16:14:22.630701747 +0000 UTC m=+154.643980314" watchObservedRunningTime="2025-09-30 16:14:22.655394493 +0000 UTC m=+154.668673020" Sep 30 16:14:22 crc kubenswrapper[4796]: I0930 16:14:22.655948 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pspfv" podStartSLOduration=2.195788905 podStartE2EDuration="54.655942658s" podCreationTimestamp="2025-09-30 16:13:28 +0000 UTC" firstStartedPulling="2025-09-30 16:13:29.862638798 +0000 UTC m=+101.875917325" lastFinishedPulling="2025-09-30 16:14:22.322792501 +0000 UTC m=+154.336071078" observedRunningTime="2025-09-30 16:14:22.652907412 +0000 UTC m=+154.666185949" watchObservedRunningTime="2025-09-30 16:14:22.655942658 +0000 UTC m=+154.669221195" Sep 30 16:14:23 crc kubenswrapper[4796]: I0930 16:14:23.246195 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-25vk7" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="registry-server" probeResult="failure" output=< Sep 30 16:14:23 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Sep 30 16:14:23 crc kubenswrapper[4796]: > Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.573615 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.574261 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.631218 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.701591 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.765609 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.765656 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:14:28 crc kubenswrapper[4796]: I0930 16:14:28.806050 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:14:29 crc kubenswrapper[4796]: I0930 16:14:29.013428 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:14:29 crc kubenswrapper[4796]: I0930 16:14:29.013591 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:14:29 crc kubenswrapper[4796]: I0930 16:14:29.082465 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:14:29 crc kubenswrapper[4796]: I0930 16:14:29.701634 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:14:29 crc kubenswrapper[4796]: I0930 16:14:29.716664 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.346028 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.346083 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.389004 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.746390 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.827107 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.867280 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:14:30 crc kubenswrapper[4796]: I0930 16:14:30.978803 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2f9rf"] Sep 30 16:14:31 crc kubenswrapper[4796]: I0930 16:14:31.097071 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:14:31 crc kubenswrapper[4796]: I0930 16:14:31.097149 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:14:31 crc kubenswrapper[4796]: I0930 16:14:31.177596 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mqzq"] Sep 30 16:14:31 crc kubenswrapper[4796]: I0930 16:14:31.673572 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4mqzq" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="registry-server" containerID="cri-o://e4c8cefed2941e393b39c68bf851c80441cd433dcba318e92b4464735f94e732" gracePeriod=2 Sep 30 16:14:32 crc kubenswrapper[4796]: I0930 16:14:32.245005 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:14:32 crc kubenswrapper[4796]: I0930 16:14:32.284794 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:14:32 crc kubenswrapper[4796]: I0930 16:14:32.685675 4796 generic.go:334] "Generic (PLEG): container finished" podID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerID="e4c8cefed2941e393b39c68bf851c80441cd433dcba318e92b4464735f94e732" exitCode=0 Sep 30 16:14:32 crc kubenswrapper[4796]: I0930 16:14:32.685764 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mqzq" event={"ID":"0b156b66-9ee3-4b88-9874-acac9e6606a1","Type":"ContainerDied","Data":"e4c8cefed2941e393b39c68bf851c80441cd433dcba318e92b4464735f94e732"} Sep 30 16:14:32 crc kubenswrapper[4796]: I0930 16:14:32.686529 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2f9rf" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="registry-server" containerID="cri-o://c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb" gracePeriod=2 Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.043538 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.228557 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-utilities\") pod \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.228624 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-catalog-content\") pod \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.228670 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktmhq\" (UniqueName: \"kubernetes.io/projected/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-kube-api-access-ktmhq\") pod \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\" (UID: \"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.229479 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-utilities" (OuterVolumeSpecName: "utilities") pod "aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" (UID: "aae79f18-b8bd-46d1-8af1-5f2acbaa66ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.235863 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-kube-api-access-ktmhq" (OuterVolumeSpecName: "kube-api-access-ktmhq") pod "aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" (UID: "aae79f18-b8bd-46d1-8af1-5f2acbaa66ab"). InnerVolumeSpecName "kube-api-access-ktmhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.283181 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" (UID: "aae79f18-b8bd-46d1-8af1-5f2acbaa66ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.321999 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.330541 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktmhq\" (UniqueName: \"kubernetes.io/projected/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-kube-api-access-ktmhq\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.330576 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.330589 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.377111 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldvs"] Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.377365 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hldvs" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="registry-server" containerID="cri-o://2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e" gracePeriod=2 Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.431972 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-utilities\") pod \"0b156b66-9ee3-4b88-9874-acac9e6606a1\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.432512 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs8sw\" (UniqueName: \"kubernetes.io/projected/0b156b66-9ee3-4b88-9874-acac9e6606a1-kube-api-access-xs8sw\") pod \"0b156b66-9ee3-4b88-9874-acac9e6606a1\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.432551 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-catalog-content\") pod \"0b156b66-9ee3-4b88-9874-acac9e6606a1\" (UID: \"0b156b66-9ee3-4b88-9874-acac9e6606a1\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.432930 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-utilities" (OuterVolumeSpecName: "utilities") pod "0b156b66-9ee3-4b88-9874-acac9e6606a1" (UID: "0b156b66-9ee3-4b88-9874-acac9e6606a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.435876 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b156b66-9ee3-4b88-9874-acac9e6606a1-kube-api-access-xs8sw" (OuterVolumeSpecName: "kube-api-access-xs8sw") pod "0b156b66-9ee3-4b88-9874-acac9e6606a1" (UID: "0b156b66-9ee3-4b88-9874-acac9e6606a1"). InnerVolumeSpecName "kube-api-access-xs8sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.489457 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b156b66-9ee3-4b88-9874-acac9e6606a1" (UID: "0b156b66-9ee3-4b88-9874-acac9e6606a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.534491 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs8sw\" (UniqueName: \"kubernetes.io/projected/0b156b66-9ee3-4b88-9874-acac9e6606a1-kube-api-access-xs8sw\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.534537 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.534549 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b156b66-9ee3-4b88-9874-acac9e6606a1-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.675696 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.716387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mqzq" event={"ID":"0b156b66-9ee3-4b88-9874-acac9e6606a1","Type":"ContainerDied","Data":"fb630c1919e851445706f027feaa5bda02a685aaeb918270b9bc7bb45c1f37e9"} Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.716436 4796 scope.go:117] "RemoveContainer" containerID="e4c8cefed2941e393b39c68bf851c80441cd433dcba318e92b4464735f94e732" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.716402 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mqzq" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.722854 4796 generic.go:334] "Generic (PLEG): container finished" podID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerID="2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e" exitCode=0 Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.722918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldvs" event={"ID":"ccbaff9e-7239-4b52-9123-fda87849ccd4","Type":"ContainerDied","Data":"2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e"} Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.722945 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldvs" event={"ID":"ccbaff9e-7239-4b52-9123-fda87849ccd4","Type":"ContainerDied","Data":"ab0da10d6343368372d53e490558dff68f2c91e5a0ccf67b356288644748b385"} Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.723033 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldvs" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.741318 4796 generic.go:334] "Generic (PLEG): container finished" podID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerID="c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb" exitCode=0 Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.741355 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9rf" event={"ID":"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab","Type":"ContainerDied","Data":"c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb"} Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.741378 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9rf" event={"ID":"aae79f18-b8bd-46d1-8af1-5f2acbaa66ab","Type":"ContainerDied","Data":"25de62b1ae802fc09e8828143514b0267b8f04222cb2292d45d7a673b7f78ede"} Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.741427 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9rf" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.771099 4796 scope.go:117] "RemoveContainer" containerID="84dc4a0c1d54ee5ab9c4960779d72e38a6fd4a82f44786d24ce2fee939475660" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.779334 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mqzq"] Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.782116 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4mqzq"] Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.789093 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2f9rf"] Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.794852 4796 scope.go:117] "RemoveContainer" containerID="65cabb7899a2728ce70424af0ebac1f48a8af14f90f30256b82be183b09d9273" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.804356 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2f9rf"] Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.809031 4796 scope.go:117] "RemoveContainer" containerID="2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.822042 4796 scope.go:117] "RemoveContainer" containerID="db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.833743 4796 scope.go:117] "RemoveContainer" containerID="7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.837757 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-catalog-content\") pod \"ccbaff9e-7239-4b52-9123-fda87849ccd4\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.837822 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-utilities\") pod \"ccbaff9e-7239-4b52-9123-fda87849ccd4\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.837905 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6s8r\" (UniqueName: \"kubernetes.io/projected/ccbaff9e-7239-4b52-9123-fda87849ccd4-kube-api-access-l6s8r\") pod \"ccbaff9e-7239-4b52-9123-fda87849ccd4\" (UID: \"ccbaff9e-7239-4b52-9123-fda87849ccd4\") " Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.839821 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-utilities" (OuterVolumeSpecName: "utilities") pod "ccbaff9e-7239-4b52-9123-fda87849ccd4" (UID: "ccbaff9e-7239-4b52-9123-fda87849ccd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.849496 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccbaff9e-7239-4b52-9123-fda87849ccd4-kube-api-access-l6s8r" (OuterVolumeSpecName: "kube-api-access-l6s8r") pod "ccbaff9e-7239-4b52-9123-fda87849ccd4" (UID: "ccbaff9e-7239-4b52-9123-fda87849ccd4"). InnerVolumeSpecName "kube-api-access-l6s8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.853072 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccbaff9e-7239-4b52-9123-fda87849ccd4" (UID: "ccbaff9e-7239-4b52-9123-fda87849ccd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.853516 4796 scope.go:117] "RemoveContainer" containerID="2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e" Sep 30 16:14:33 crc kubenswrapper[4796]: E0930 16:14:33.854068 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e\": container with ID starting with 2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e not found: ID does not exist" containerID="2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.854112 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e"} err="failed to get container status \"2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e\": rpc error: code = NotFound desc = could not find container \"2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e\": container with ID starting with 2df0d5e70b01afb156206b90f5ee04bda93132d3989a3e64e365a2fe3824f48e not found: ID does not exist" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.854198 4796 scope.go:117] "RemoveContainer" containerID="db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438" Sep 30 16:14:33 crc kubenswrapper[4796]: E0930 16:14:33.861888 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438\": container with ID starting with db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438 not found: ID does not exist" containerID="db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.862029 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438"} err="failed to get container status \"db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438\": rpc error: code = NotFound desc = could not find container \"db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438\": container with ID starting with db950094c9fb9556df5791cbb66d862385f1feccd18f9b9ef0626658b85ce438 not found: ID does not exist" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.862128 4796 scope.go:117] "RemoveContainer" containerID="7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04" Sep 30 16:14:33 crc kubenswrapper[4796]: E0930 16:14:33.862632 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04\": container with ID starting with 7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04 not found: ID does not exist" containerID="7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.862733 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04"} err="failed to get container status \"7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04\": rpc error: code = NotFound desc = could not find container \"7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04\": container with ID starting with 7ff548d33e12c5985662d75f6af3874b68b845fcff79a4c97100cf20080e8c04 not found: ID does not exist" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.862805 4796 scope.go:117] "RemoveContainer" containerID="c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.874935 4796 scope.go:117] "RemoveContainer" containerID="7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.888152 4796 scope.go:117] "RemoveContainer" containerID="0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.900333 4796 scope.go:117] "RemoveContainer" containerID="c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb" Sep 30 16:14:33 crc kubenswrapper[4796]: E0930 16:14:33.900697 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb\": container with ID starting with c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb not found: ID does not exist" containerID="c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.900816 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb"} err="failed to get container status \"c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb\": rpc error: code = NotFound desc = could not find container \"c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb\": container with ID starting with c56038cdba22e0164014c2263e1c421ea879b223654dfc4fdaed5530403d0bdb not found: ID does not exist" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.900916 4796 scope.go:117] "RemoveContainer" containerID="7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3" Sep 30 16:14:33 crc kubenswrapper[4796]: E0930 16:14:33.901219 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3\": container with ID starting with 7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3 not found: ID does not exist" containerID="7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.901298 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3"} err="failed to get container status \"7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3\": rpc error: code = NotFound desc = could not find container \"7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3\": container with ID starting with 7bafda882a6fbbfe4862e8db043f14ff39f0733b19165fa98c8fec5392e8f7f3 not found: ID does not exist" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.901372 4796 scope.go:117] "RemoveContainer" containerID="0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c" Sep 30 16:14:33 crc kubenswrapper[4796]: E0930 16:14:33.901660 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c\": container with ID starting with 0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c not found: ID does not exist" containerID="0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.901788 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c"} err="failed to get container status \"0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c\": rpc error: code = NotFound desc = could not find container \"0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c\": container with ID starting with 0a3228b7c2e5705cf9a12f60f6351cfa42bb2989953df41e06175b4afe164c4c not found: ID does not exist" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.940099 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.940495 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6s8r\" (UniqueName: \"kubernetes.io/projected/ccbaff9e-7239-4b52-9123-fda87849ccd4-kube-api-access-l6s8r\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:33 crc kubenswrapper[4796]: I0930 16:14:33.940611 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccbaff9e-7239-4b52-9123-fda87849ccd4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:34 crc kubenswrapper[4796]: I0930 16:14:34.053893 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldvs"] Sep 30 16:14:34 crc kubenswrapper[4796]: I0930 16:14:34.056936 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldvs"] Sep 30 16:14:34 crc kubenswrapper[4796]: I0930 16:14:34.740786 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" path="/var/lib/kubelet/pods/0b156b66-9ee3-4b88-9874-acac9e6606a1/volumes" Sep 30 16:14:34 crc kubenswrapper[4796]: I0930 16:14:34.741893 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" path="/var/lib/kubelet/pods/aae79f18-b8bd-46d1-8af1-5f2acbaa66ab/volumes" Sep 30 16:14:34 crc kubenswrapper[4796]: I0930 16:14:34.742452 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" path="/var/lib/kubelet/pods/ccbaff9e-7239-4b52-9123-fda87849ccd4/volumes" Sep 30 16:14:35 crc kubenswrapper[4796]: I0930 16:14:35.778679 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25vk7"] Sep 30 16:14:35 crc kubenswrapper[4796]: I0930 16:14:35.778988 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-25vk7" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="registry-server" containerID="cri-o://ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0" gracePeriod=2 Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.140208 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.267553 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-catalog-content\") pod \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.267704 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-utilities\") pod \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.267750 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g6ft\" (UniqueName: \"kubernetes.io/projected/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-kube-api-access-4g6ft\") pod \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\" (UID: \"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075\") " Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.268811 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-utilities" (OuterVolumeSpecName: "utilities") pod "8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" (UID: "8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.274335 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-kube-api-access-4g6ft" (OuterVolumeSpecName: "kube-api-access-4g6ft") pod "8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" (UID: "8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075"). InnerVolumeSpecName "kube-api-access-4g6ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.357559 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" (UID: "8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.369560 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.369593 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g6ft\" (UniqueName: \"kubernetes.io/projected/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-kube-api-access-4g6ft\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.369629 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.769566 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerID="ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0" exitCode=0 Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.769620 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25vk7" event={"ID":"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075","Type":"ContainerDied","Data":"ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0"} Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.769641 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25vk7" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.769656 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25vk7" event={"ID":"8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075","Type":"ContainerDied","Data":"f20fcbf64403d082cf4b930044c56af45dfaa819fe715b7762e7d5defa2a1aeb"} Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.769680 4796 scope.go:117] "RemoveContainer" containerID="ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.786401 4796 scope.go:117] "RemoveContainer" containerID="be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.792080 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25vk7"] Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.794603 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-25vk7"] Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.807683 4796 scope.go:117] "RemoveContainer" containerID="b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.826706 4796 scope.go:117] "RemoveContainer" containerID="ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0" Sep 30 16:14:36 crc kubenswrapper[4796]: E0930 16:14:36.829363 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0\": container with ID starting with ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0 not found: ID does not exist" containerID="ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.829426 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0"} err="failed to get container status \"ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0\": rpc error: code = NotFound desc = could not find container \"ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0\": container with ID starting with ae0d0609c5f0ef1644fa34fc778b5d3ba4e6e72b8a5ec5caaa172d1a5f53f3e0 not found: ID does not exist" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.829469 4796 scope.go:117] "RemoveContainer" containerID="be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc" Sep 30 16:14:36 crc kubenswrapper[4796]: E0930 16:14:36.829962 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc\": container with ID starting with be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc not found: ID does not exist" containerID="be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.830117 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc"} err="failed to get container status \"be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc\": rpc error: code = NotFound desc = could not find container \"be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc\": container with ID starting with be3243907b07e0deffaa059004fd3442710c0535568d60550569e2f6583b27dc not found: ID does not exist" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.830149 4796 scope.go:117] "RemoveContainer" containerID="b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc" Sep 30 16:14:36 crc kubenswrapper[4796]: E0930 16:14:36.830452 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc\": container with ID starting with b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc not found: ID does not exist" containerID="b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc" Sep 30 16:14:36 crc kubenswrapper[4796]: I0930 16:14:36.830498 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc"} err="failed to get container status \"b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc\": rpc error: code = NotFound desc = could not find container \"b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc\": container with ID starting with b9a5741cfc511a1c63c4d13a6f56f923c829b6e835c5d5ca6deb35455865f8bc not found: ID does not exist" Sep 30 16:14:37 crc kubenswrapper[4796]: I0930 16:14:37.743273 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mqjb"] Sep 30 16:14:38 crc kubenswrapper[4796]: I0930 16:14:38.739432 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" path="/var/lib/kubelet/pods/8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075/volumes" Sep 30 16:14:55 crc kubenswrapper[4796]: I0930 16:14:55.711374 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151024 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59"] Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151750 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151772 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151791 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151803 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151820 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151833 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151848 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05657c8-ad27-4b34-96e6-c41462e82358" containerName="pruner" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151859 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05657c8-ad27-4b34-96e6-c41462e82358" containerName="pruner" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151874 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151886 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151900 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151912 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151932 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151943 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.151960 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.151972 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.152020 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152056 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.152069 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152082 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.152100 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152115 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.152136 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152148 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="extract-content" Sep 30 16:15:00 crc kubenswrapper[4796]: E0930 16:15:00.152166 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152178 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="extract-utilities" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152347 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae79f18-b8bd-46d1-8af1-5f2acbaa66ab" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152366 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f1cc7b3-a1ac-4e29-ad9c-7af26e2e7075" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152383 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b156b66-9ee3-4b88-9874-acac9e6606a1" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152398 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f05657c8-ad27-4b34-96e6-c41462e82358" containerName="pruner" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.152416 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbaff9e-7239-4b52-9123-fda87849ccd4" containerName="registry-server" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.153061 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.155807 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.156222 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.165374 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59"] Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.280778 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvh4l\" (UniqueName: \"kubernetes.io/projected/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-kube-api-access-hvh4l\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.280940 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-secret-volume\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.281477 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-config-volume\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.384611 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-config-volume\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.384736 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvh4l\" (UniqueName: \"kubernetes.io/projected/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-kube-api-access-hvh4l\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.384807 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-secret-volume\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.386171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-config-volume\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.394015 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-secret-volume\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.404442 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvh4l\" (UniqueName: \"kubernetes.io/projected/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-kube-api-access-hvh4l\") pod \"collect-profiles-29320815-bsb59\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.491893 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:00 crc kubenswrapper[4796]: I0930 16:15:00.908163 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59"] Sep 30 16:15:01 crc kubenswrapper[4796]: I0930 16:15:01.097452 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:15:01 crc kubenswrapper[4796]: I0930 16:15:01.097929 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:15:01 crc kubenswrapper[4796]: I0930 16:15:01.906958 4796 generic.go:334] "Generic (PLEG): container finished" podID="b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" containerID="04ce5f567fa6307b9536918267c255b9b7cfca95a8e0ef1e694e35f8607eab4f" exitCode=0 Sep 30 16:15:01 crc kubenswrapper[4796]: I0930 16:15:01.907047 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" event={"ID":"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5","Type":"ContainerDied","Data":"04ce5f567fa6307b9536918267c255b9b7cfca95a8e0ef1e694e35f8607eab4f"} Sep 30 16:15:01 crc kubenswrapper[4796]: I0930 16:15:01.907096 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" event={"ID":"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5","Type":"ContainerStarted","Data":"361c00c403ce75d26744f6c257ccac871e759f78d0f350bb31f40583faed5f33"} Sep 30 16:15:02 crc kubenswrapper[4796]: I0930 16:15:02.771081 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" podUID="2434945c-60ef-40c1-aa33-2cded03cc0e0" containerName="oauth-openshift" containerID="cri-o://deae6dfad6e2e288357909a067910f107093a549fc8925bc1f521a80aa440162" gracePeriod=15 Sep 30 16:15:02 crc kubenswrapper[4796]: I0930 16:15:02.923070 4796 generic.go:334] "Generic (PLEG): container finished" podID="2434945c-60ef-40c1-aa33-2cded03cc0e0" containerID="deae6dfad6e2e288357909a067910f107093a549fc8925bc1f521a80aa440162" exitCode=0 Sep 30 16:15:02 crc kubenswrapper[4796]: I0930 16:15:02.923191 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" event={"ID":"2434945c-60ef-40c1-aa33-2cded03cc0e0","Type":"ContainerDied","Data":"deae6dfad6e2e288357909a067910f107093a549fc8925bc1f521a80aa440162"} Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.214620 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.221018 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.221849 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-session\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.221895 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-ocp-branding-template\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.221922 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-router-certs\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.221956 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-service-ca\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.222327 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.222903 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223127 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-dir\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223180 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-cliconfig\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223213 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-login\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223240 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-trusted-ca-bundle\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223267 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52l66\" (UniqueName: \"kubernetes.io/projected/2434945c-60ef-40c1-aa33-2cded03cc0e0-kube-api-access-52l66\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223297 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-idp-0-file-data\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223335 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-provider-selection\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223370 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-serving-cert\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223454 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-error\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223478 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-policies\") pod \"2434945c-60ef-40c1-aa33-2cded03cc0e0\" (UID: \"2434945c-60ef-40c1-aa33-2cded03cc0e0\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223760 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.223778 4796 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.224277 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.224802 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.225429 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.229835 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.230080 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.230637 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.230710 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2434945c-60ef-40c1-aa33-2cded03cc0e0-kube-api-access-52l66" (OuterVolumeSpecName: "kube-api-access-52l66") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "kube-api-access-52l66". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.231396 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.231522 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.232204 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.233142 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.233458 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2434945c-60ef-40c1-aa33-2cded03cc0e0" (UID: "2434945c-60ef-40c1-aa33-2cded03cc0e0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.253663 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-cc7989dc6-glnqf"] Sep 30 16:15:03 crc kubenswrapper[4796]: E0930 16:15:03.253843 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2434945c-60ef-40c1-aa33-2cded03cc0e0" containerName="oauth-openshift" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.253854 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2434945c-60ef-40c1-aa33-2cded03cc0e0" containerName="oauth-openshift" Sep 30 16:15:03 crc kubenswrapper[4796]: E0930 16:15:03.253868 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" containerName="collect-profiles" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.253877 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" containerName="collect-profiles" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.254000 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" containerName="collect-profiles" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.254017 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2434945c-60ef-40c1-aa33-2cded03cc0e0" containerName="oauth-openshift" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.254422 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.277250 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-cc7989dc6-glnqf"] Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.324745 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-secret-volume\") pod \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.324812 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-config-volume\") pod \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.324878 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvh4l\" (UniqueName: \"kubernetes.io/projected/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-kube-api-access-hvh4l\") pod \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\" (UID: \"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5\") " Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.324966 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-audit-policies\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325012 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325033 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-session\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325049 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-login\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325066 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-service-ca\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325088 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325115 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325130 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325149 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-audit-dir\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325182 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-error\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325200 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325220 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-router-certs\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325235 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd69w\" (UniqueName: \"kubernetes.io/projected/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-kube-api-access-kd69w\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325271 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325280 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325289 4796 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325298 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325309 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325320 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325329 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325337 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325346 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325356 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52l66\" (UniqueName: \"kubernetes.io/projected/2434945c-60ef-40c1-aa33-2cded03cc0e0-kube-api-access-52l66\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325365 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.325374 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2434945c-60ef-40c1-aa33-2cded03cc0e0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.326687 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-config-volume" (OuterVolumeSpecName: "config-volume") pod "b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" (UID: "b0e43d3f-51c9-4caf-a2b4-394b0f4729f5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.328923 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-kube-api-access-hvh4l" (OuterVolumeSpecName: "kube-api-access-hvh4l") pod "b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" (UID: "b0e43d3f-51c9-4caf-a2b4-394b0f4729f5"). InnerVolumeSpecName "kube-api-access-hvh4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.329164 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" (UID: "b0e43d3f-51c9-4caf-a2b4-394b0f4729f5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427175 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427306 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427367 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427408 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-audit-dir\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-error\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427488 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427526 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-router-certs\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427562 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd69w\" (UniqueName: \"kubernetes.io/projected/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-kube-api-access-kd69w\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427608 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-audit-policies\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427630 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-audit-dir\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427702 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427771 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-session\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427810 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-login\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427862 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-service-ca\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427949 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvh4l\" (UniqueName: \"kubernetes.io/projected/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-kube-api-access-hvh4l\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.427972 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.428027 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.428396 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.429541 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.429642 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-service-ca\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.429824 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-audit-policies\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.432971 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-error\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.433651 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.434108 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.434540 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.436291 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.436134 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-router-certs\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.437925 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-user-template-login\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.438992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-v4-0-config-system-session\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.448100 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd69w\" (UniqueName: \"kubernetes.io/projected/8c20def2-9b46-4373-a9cb-6a22b7db7bc9-kube-api-access-kd69w\") pod \"oauth-openshift-cc7989dc6-glnqf\" (UID: \"8c20def2-9b46-4373-a9cb-6a22b7db7bc9\") " pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.591531 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.810084 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-cc7989dc6-glnqf"] Sep 30 16:15:03 crc kubenswrapper[4796]: W0930 16:15:03.815835 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c20def2_9b46_4373_a9cb_6a22b7db7bc9.slice/crio-e25814da3b806a42e9f1d20b6b8f0c68c080c00887a45f72c7962aacf305a0c8 WatchSource:0}: Error finding container e25814da3b806a42e9f1d20b6b8f0c68c080c00887a45f72c7962aacf305a0c8: Status 404 returned error can't find the container with id e25814da3b806a42e9f1d20b6b8f0c68c080c00887a45f72c7962aacf305a0c8 Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.933340 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" event={"ID":"b0e43d3f-51c9-4caf-a2b4-394b0f4729f5","Type":"ContainerDied","Data":"361c00c403ce75d26744f6c257ccac871e759f78d0f350bb31f40583faed5f33"} Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.933393 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="361c00c403ce75d26744f6c257ccac871e759f78d0f350bb31f40583faed5f33" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.933396 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.935646 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.936442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mqjb" event={"ID":"2434945c-60ef-40c1-aa33-2cded03cc0e0","Type":"ContainerDied","Data":"694dba313a4a204045eb8dd2ee43659fd696a153c497e83a99d0ad93fec3d0b9"} Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.936508 4796 scope.go:117] "RemoveContainer" containerID="deae6dfad6e2e288357909a067910f107093a549fc8925bc1f521a80aa440162" Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.937836 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" event={"ID":"8c20def2-9b46-4373-a9cb-6a22b7db7bc9","Type":"ContainerStarted","Data":"e25814da3b806a42e9f1d20b6b8f0c68c080c00887a45f72c7962aacf305a0c8"} Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.979840 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mqjb"] Sep 30 16:15:03 crc kubenswrapper[4796]: I0930 16:15:03.983293 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mqjb"] Sep 30 16:15:04 crc kubenswrapper[4796]: I0930 16:15:04.754637 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2434945c-60ef-40c1-aa33-2cded03cc0e0" path="/var/lib/kubelet/pods/2434945c-60ef-40c1-aa33-2cded03cc0e0/volumes" Sep 30 16:15:04 crc kubenswrapper[4796]: I0930 16:15:04.953568 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" event={"ID":"8c20def2-9b46-4373-a9cb-6a22b7db7bc9","Type":"ContainerStarted","Data":"0ebf8a943cf0c184f2329e222b90882152e99fceb8d9c8841bb7b442fa4c3ce0"} Sep 30 16:15:04 crc kubenswrapper[4796]: I0930 16:15:04.954022 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:04 crc kubenswrapper[4796]: I0930 16:15:04.963372 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" Sep 30 16:15:04 crc kubenswrapper[4796]: I0930 16:15:04.979152 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-cc7989dc6-glnqf" podStartSLOduration=27.979091663 podStartE2EDuration="27.979091663s" podCreationTimestamp="2025-09-30 16:14:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:15:04.977193289 +0000 UTC m=+196.990471836" watchObservedRunningTime="2025-09-30 16:15:04.979091663 +0000 UTC m=+196.992370190" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.436377 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgjk8"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.437613 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wgjk8" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="registry-server" containerID="cri-o://5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f" gracePeriod=30 Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.444615 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pspfv"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.444882 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pspfv" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="registry-server" containerID="cri-o://a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486" gracePeriod=30 Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.455285 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbg46"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.455721 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerName="marketplace-operator" containerID="cri-o://6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193" gracePeriod=30 Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.478806 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhk59"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.479236 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lhk59" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="registry-server" containerID="cri-o://008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060" gracePeriod=30 Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.482838 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lchcj"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.483539 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.489085 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8xs5d"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.489304 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8xs5d" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="registry-server" containerID="cri-o://762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565" gracePeriod=30 Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.491601 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lchcj"] Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.615285 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlf96\" (UniqueName: \"kubernetes.io/projected/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-kube-api-access-dlf96\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.615364 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.615401 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.716870 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.717361 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlf96\" (UniqueName: \"kubernetes.io/projected/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-kube-api-access-dlf96\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.717384 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.719063 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.725962 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.738913 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlf96\" (UniqueName: \"kubernetes.io/projected/f6ab27c4-6027-4d97-88c0-a8ad5eac035a-kube-api-access-dlf96\") pod \"marketplace-operator-79b997595-lchcj\" (UID: \"f6ab27c4-6027-4d97-88c0-a8ad5eac035a\") " pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.858604 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.868820 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.869188 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.882614 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.901883 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:15:17 crc kubenswrapper[4796]: I0930 16:15:17.923179 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.020890 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-catalog-content\") pod \"c0e5842f-b782-477d-87d7-c677e786c155\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.020951 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skfm9\" (UniqueName: \"kubernetes.io/projected/25f6d193-b5f7-49cc-acf0-1698e5a15de7-kube-api-access-skfm9\") pod \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021023 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-utilities\") pod \"c0e5842f-b782-477d-87d7-c677e786c155\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021048 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-catalog-content\") pod \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021076 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-trusted-ca\") pod \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021107 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-catalog-content\") pod \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021158 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-operator-metrics\") pod \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021181 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfbs6\" (UniqueName: \"kubernetes.io/projected/78afec2a-ad1c-4999-8165-eda2ed82ef5d-kube-api-access-xfbs6\") pod \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021219 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-utilities\") pod \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\" (UID: \"25f6d193-b5f7-49cc-acf0-1698e5a15de7\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021271 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-utilities\") pod \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\" (UID: \"78afec2a-ad1c-4999-8165-eda2ed82ef5d\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021294 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55sc4\" (UniqueName: \"kubernetes.io/projected/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-kube-api-access-55sc4\") pod \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\" (UID: \"7881eaa3-77bc-430c-a2a6-d44ada95cc0a\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.021315 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fplfk\" (UniqueName: \"kubernetes.io/projected/c0e5842f-b782-477d-87d7-c677e786c155-kube-api-access-fplfk\") pod \"c0e5842f-b782-477d-87d7-c677e786c155\" (UID: \"c0e5842f-b782-477d-87d7-c677e786c155\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.022301 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-utilities" (OuterVolumeSpecName: "utilities") pod "c0e5842f-b782-477d-87d7-c677e786c155" (UID: "c0e5842f-b782-477d-87d7-c677e786c155"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.024036 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7881eaa3-77bc-430c-a2a6-d44ada95cc0a" (UID: "7881eaa3-77bc-430c-a2a6-d44ada95cc0a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.027113 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-utilities" (OuterVolumeSpecName: "utilities") pod "78afec2a-ad1c-4999-8165-eda2ed82ef5d" (UID: "78afec2a-ad1c-4999-8165-eda2ed82ef5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.029778 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-utilities" (OuterVolumeSpecName: "utilities") pod "25f6d193-b5f7-49cc-acf0-1698e5a15de7" (UID: "25f6d193-b5f7-49cc-acf0-1698e5a15de7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.031996 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f6d193-b5f7-49cc-acf0-1698e5a15de7-kube-api-access-skfm9" (OuterVolumeSpecName: "kube-api-access-skfm9") pod "25f6d193-b5f7-49cc-acf0-1698e5a15de7" (UID: "25f6d193-b5f7-49cc-acf0-1698e5a15de7"). InnerVolumeSpecName "kube-api-access-skfm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.032024 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78afec2a-ad1c-4999-8165-eda2ed82ef5d-kube-api-access-xfbs6" (OuterVolumeSpecName: "kube-api-access-xfbs6") pod "78afec2a-ad1c-4999-8165-eda2ed82ef5d" (UID: "78afec2a-ad1c-4999-8165-eda2ed82ef5d"). InnerVolumeSpecName "kube-api-access-xfbs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.032224 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0e5842f-b782-477d-87d7-c677e786c155-kube-api-access-fplfk" (OuterVolumeSpecName: "kube-api-access-fplfk") pod "c0e5842f-b782-477d-87d7-c677e786c155" (UID: "c0e5842f-b782-477d-87d7-c677e786c155"). InnerVolumeSpecName "kube-api-access-fplfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.032226 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7881eaa3-77bc-430c-a2a6-d44ada95cc0a" (UID: "7881eaa3-77bc-430c-a2a6-d44ada95cc0a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.038929 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25f6d193-b5f7-49cc-acf0-1698e5a15de7" (UID: "25f6d193-b5f7-49cc-acf0-1698e5a15de7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.039915 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-kube-api-access-55sc4" (OuterVolumeSpecName: "kube-api-access-55sc4") pod "7881eaa3-77bc-430c-a2a6-d44ada95cc0a" (UID: "7881eaa3-77bc-430c-a2a6-d44ada95cc0a"). InnerVolumeSpecName "kube-api-access-55sc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.055763 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.055763 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" event={"ID":"7881eaa3-77bc-430c-a2a6-d44ada95cc0a","Type":"ContainerDied","Data":"6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.055840 4796 scope.go:117] "RemoveContainer" containerID="6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.055627 4796 generic.go:334] "Generic (PLEG): container finished" podID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerID="6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193" exitCode=0 Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.069207 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bbg46" event={"ID":"7881eaa3-77bc-430c-a2a6-d44ada95cc0a","Type":"ContainerDied","Data":"fa77db16d03c19e95957e4fe8ab4d72dbefaf8b99a5b40548b7d3dd036f0739b"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.075560 4796 generic.go:334] "Generic (PLEG): container finished" podID="c0e5842f-b782-477d-87d7-c677e786c155" containerID="5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f" exitCode=0 Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.075621 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgjk8" event={"ID":"c0e5842f-b782-477d-87d7-c677e786c155","Type":"ContainerDied","Data":"5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.075594 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgjk8" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.075676 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgjk8" event={"ID":"c0e5842f-b782-477d-87d7-c677e786c155","Type":"ContainerDied","Data":"7edd1acf1a4ee10ad6f189de9483dc74bcb7be75a84b9b1036ea3bdd4b390414"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.093433 4796 scope.go:117] "RemoveContainer" containerID="6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.093799 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193\": container with ID starting with 6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193 not found: ID does not exist" containerID="6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.093848 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193"} err="failed to get container status \"6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193\": rpc error: code = NotFound desc = could not find container \"6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193\": container with ID starting with 6bf6e13fb812ccbfa69147c0aa03fb3c0b2bd9983b9c66131d6f229a0371a193 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.093938 4796 scope.go:117] "RemoveContainer" containerID="5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.095215 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78afec2a-ad1c-4999-8165-eda2ed82ef5d" (UID: "78afec2a-ad1c-4999-8165-eda2ed82ef5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.096703 4796 generic.go:334] "Generic (PLEG): container finished" podID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerID="762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565" exitCode=0 Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.096805 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xs5d" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.097031 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbg46"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.097073 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerDied","Data":"762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.097102 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xs5d" event={"ID":"a77743d1-a6ef-4741-83d2-a102d2a256dc","Type":"ContainerDied","Data":"4851808ac3ef5e9980c59344d82c4ae128766562cb9a4d8f445fd7ffce135962"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.099921 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbg46"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.102868 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0e5842f-b782-477d-87d7-c677e786c155" (UID: "c0e5842f-b782-477d-87d7-c677e786c155"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.108501 4796 generic.go:334] "Generic (PLEG): container finished" podID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerID="a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486" exitCode=0 Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.108614 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerDied","Data":"a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.108643 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pspfv" event={"ID":"78afec2a-ad1c-4999-8165-eda2ed82ef5d","Type":"ContainerDied","Data":"e8c11238ec3be1f9f8a001af598414cd8514c1596908bf48ff01987350197d27"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.108907 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pspfv" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.112397 4796 generic.go:334] "Generic (PLEG): container finished" podID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerID="008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060" exitCode=0 Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.112424 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhk59" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.112442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhk59" event={"ID":"25f6d193-b5f7-49cc-acf0-1698e5a15de7","Type":"ContainerDied","Data":"008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.112795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhk59" event={"ID":"25f6d193-b5f7-49cc-acf0-1698e5a15de7","Type":"ContainerDied","Data":"584323153f468cb319e1939c4420075d6e0f7e8c4bdfd3ebdc073c0c92c7e719"} Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.122066 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-catalog-content\") pod \"a77743d1-a6ef-4741-83d2-a102d2a256dc\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.122263 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st2ww\" (UniqueName: \"kubernetes.io/projected/a77743d1-a6ef-4741-83d2-a102d2a256dc-kube-api-access-st2ww\") pod \"a77743d1-a6ef-4741-83d2-a102d2a256dc\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.122367 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-utilities\") pod \"a77743d1-a6ef-4741-83d2-a102d2a256dc\" (UID: \"a77743d1-a6ef-4741-83d2-a102d2a256dc\") " Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.122756 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.122796 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfbs6\" (UniqueName: \"kubernetes.io/projected/78afec2a-ad1c-4999-8165-eda2ed82ef5d-kube-api-access-xfbs6\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.122866 4796 scope.go:117] "RemoveContainer" containerID="4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123190 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123222 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123233 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fplfk\" (UniqueName: \"kubernetes.io/projected/c0e5842f-b782-477d-87d7-c677e786c155-kube-api-access-fplfk\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123242 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55sc4\" (UniqueName: \"kubernetes.io/projected/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-kube-api-access-55sc4\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123251 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123260 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skfm9\" (UniqueName: \"kubernetes.io/projected/25f6d193-b5f7-49cc-acf0-1698e5a15de7-kube-api-access-skfm9\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123268 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f6d193-b5f7-49cc-acf0-1698e5a15de7-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123277 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e5842f-b782-477d-87d7-c677e786c155-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123287 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7881eaa3-77bc-430c-a2a6-d44ada95cc0a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123295 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78afec2a-ad1c-4999-8165-eda2ed82ef5d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.123553 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-utilities" (OuterVolumeSpecName: "utilities") pod "a77743d1-a6ef-4741-83d2-a102d2a256dc" (UID: "a77743d1-a6ef-4741-83d2-a102d2a256dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.124759 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lchcj"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.128670 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a77743d1-a6ef-4741-83d2-a102d2a256dc-kube-api-access-st2ww" (OuterVolumeSpecName: "kube-api-access-st2ww") pod "a77743d1-a6ef-4741-83d2-a102d2a256dc" (UID: "a77743d1-a6ef-4741-83d2-a102d2a256dc"). InnerVolumeSpecName "kube-api-access-st2ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.148497 4796 scope.go:117] "RemoveContainer" containerID="a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.156719 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhk59"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.165152 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhk59"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.168197 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pspfv"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.171253 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pspfv"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.181657 4796 scope.go:117] "RemoveContainer" containerID="5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.182315 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f\": container with ID starting with 5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f not found: ID does not exist" containerID="5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.182361 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f"} err="failed to get container status \"5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f\": rpc error: code = NotFound desc = could not find container \"5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f\": container with ID starting with 5db82c311be0d2c589ef581c86d96e370d2cbad0a20e276a9975f689a4bb253f not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.182392 4796 scope.go:117] "RemoveContainer" containerID="4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.182673 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b\": container with ID starting with 4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b not found: ID does not exist" containerID="4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.182725 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b"} err="failed to get container status \"4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b\": rpc error: code = NotFound desc = could not find container \"4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b\": container with ID starting with 4f16002580ddf337bcc1aa36d6fa1a03c02e0ce9106b1584e9e0c67ec9308c1b not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.182746 4796 scope.go:117] "RemoveContainer" containerID="a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.183026 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd\": container with ID starting with a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd not found: ID does not exist" containerID="a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.183055 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd"} err="failed to get container status \"a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd\": rpc error: code = NotFound desc = could not find container \"a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd\": container with ID starting with a60233a1c9484e70b673d27b271ac89060e2a2030fa268c0aa4a893825af72cd not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.183072 4796 scope.go:117] "RemoveContainer" containerID="762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.211207 4796 scope.go:117] "RemoveContainer" containerID="a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.224229 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st2ww\" (UniqueName: \"kubernetes.io/projected/a77743d1-a6ef-4741-83d2-a102d2a256dc-kube-api-access-st2ww\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.224253 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.236471 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a77743d1-a6ef-4741-83d2-a102d2a256dc" (UID: "a77743d1-a6ef-4741-83d2-a102d2a256dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.239064 4796 scope.go:117] "RemoveContainer" containerID="4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.253628 4796 scope.go:117] "RemoveContainer" containerID="762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.254024 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565\": container with ID starting with 762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565 not found: ID does not exist" containerID="762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.254063 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565"} err="failed to get container status \"762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565\": rpc error: code = NotFound desc = could not find container \"762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565\": container with ID starting with 762e21c33b4e8854f383821d959f4da018faf6a59740bf9183dccabe3c341565 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.254093 4796 scope.go:117] "RemoveContainer" containerID="a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.254448 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185\": container with ID starting with a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185 not found: ID does not exist" containerID="a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.254500 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185"} err="failed to get container status \"a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185\": rpc error: code = NotFound desc = could not find container \"a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185\": container with ID starting with a12bde53cdb9802382d9ad44d4d060597f1b0fdcf52fa1ea8e8e658050e78185 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.254536 4796 scope.go:117] "RemoveContainer" containerID="4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.254917 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26\": container with ID starting with 4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26 not found: ID does not exist" containerID="4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.254945 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26"} err="failed to get container status \"4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26\": rpc error: code = NotFound desc = could not find container \"4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26\": container with ID starting with 4f58ea46c7fe0ef4cb7bfba62e46ac63b093614de37435ecab8c176a5315ef26 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.254963 4796 scope.go:117] "RemoveContainer" containerID="a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.268859 4796 scope.go:117] "RemoveContainer" containerID="9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.284216 4796 scope.go:117] "RemoveContainer" containerID="1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.297341 4796 scope.go:117] "RemoveContainer" containerID="a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.298032 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486\": container with ID starting with a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486 not found: ID does not exist" containerID="a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.298222 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486"} err="failed to get container status \"a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486\": rpc error: code = NotFound desc = could not find container \"a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486\": container with ID starting with a66e481e3bdca4a563471a59c8fd7aa7649b24b2b84cfe74337055ba5bb83486 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.298306 4796 scope.go:117] "RemoveContainer" containerID="9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.299463 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17\": container with ID starting with 9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17 not found: ID does not exist" containerID="9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.299564 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17"} err="failed to get container status \"9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17\": rpc error: code = NotFound desc = could not find container \"9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17\": container with ID starting with 9fa2ab81eb64441da71279e9bde6a8dd3323ae79f4d937a109ef533c1a6f3c17 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.299623 4796 scope.go:117] "RemoveContainer" containerID="1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.300046 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71\": container with ID starting with 1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71 not found: ID does not exist" containerID="1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.300105 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71"} err="failed to get container status \"1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71\": rpc error: code = NotFound desc = could not find container \"1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71\": container with ID starting with 1ee06291e23548acbf507464a90afdf355c1411830120648bc524294b10f5d71 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.300142 4796 scope.go:117] "RemoveContainer" containerID="008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.317591 4796 scope.go:117] "RemoveContainer" containerID="35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.325502 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77743d1-a6ef-4741-83d2-a102d2a256dc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.330689 4796 scope.go:117] "RemoveContainer" containerID="7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.352017 4796 scope.go:117] "RemoveContainer" containerID="008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.352511 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060\": container with ID starting with 008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060 not found: ID does not exist" containerID="008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.352547 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060"} err="failed to get container status \"008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060\": rpc error: code = NotFound desc = could not find container \"008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060\": container with ID starting with 008221001fd494853d53eab3a22fef4a74992353459cba2f3557b292ded30060 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.352575 4796 scope.go:117] "RemoveContainer" containerID="35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.352911 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e\": container with ID starting with 35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e not found: ID does not exist" containerID="35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.352950 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e"} err="failed to get container status \"35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e\": rpc error: code = NotFound desc = could not find container \"35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e\": container with ID starting with 35afd21c8ea1799e6a559fc90c81395354c9734652f2e707ccc380ec3890460e not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.353006 4796 scope.go:117] "RemoveContainer" containerID="7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21" Sep 30 16:15:18 crc kubenswrapper[4796]: E0930 16:15:18.353337 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21\": container with ID starting with 7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21 not found: ID does not exist" containerID="7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.353363 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21"} err="failed to get container status \"7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21\": rpc error: code = NotFound desc = could not find container \"7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21\": container with ID starting with 7496eeed7ffbe79ee517f8a512b07cf2ea778047ff6bccd16b42b940af293c21 not found: ID does not exist" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.403806 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgjk8"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.407791 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wgjk8"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.420032 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8xs5d"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.423739 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8xs5d"] Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.741683 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" path="/var/lib/kubelet/pods/25f6d193-b5f7-49cc-acf0-1698e5a15de7/volumes" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.742726 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" path="/var/lib/kubelet/pods/7881eaa3-77bc-430c-a2a6-d44ada95cc0a/volumes" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.743447 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" path="/var/lib/kubelet/pods/78afec2a-ad1c-4999-8165-eda2ed82ef5d/volumes" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.744847 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" path="/var/lib/kubelet/pods/a77743d1-a6ef-4741-83d2-a102d2a256dc/volumes" Sep 30 16:15:18 crc kubenswrapper[4796]: I0930 16:15:18.745712 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0e5842f-b782-477d-87d7-c677e786c155" path="/var/lib/kubelet/pods/c0e5842f-b782-477d-87d7-c677e786c155/volumes" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.118348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" event={"ID":"f6ab27c4-6027-4d97-88c0-a8ad5eac035a","Type":"ContainerStarted","Data":"2de93f86b9b37179ed55e31362fda80f52b47d70931b33bfe6f5dc7b30af295c"} Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.119085 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" event={"ID":"f6ab27c4-6027-4d97-88c0-a8ad5eac035a","Type":"ContainerStarted","Data":"807b442fc63f6c20ae8dfa381c1dbfe77296af8c2f2b003f186d08ce7422cb0f"} Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.119139 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.124633 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.145348 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-lchcj" podStartSLOduration=2.145326919 podStartE2EDuration="2.145326919s" podCreationTimestamp="2025-09-30 16:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:15:19.138385442 +0000 UTC m=+211.151663969" watchObservedRunningTime="2025-09-30 16:15:19.145326919 +0000 UTC m=+211.158605446" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663013 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n8r5j"] Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663284 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663302 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663317 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663327 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663348 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663359 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663373 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663383 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663392 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663402 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663417 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663426 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663439 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663449 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663467 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663478 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663490 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663499 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="extract-content" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663511 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663521 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663535 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663545 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663558 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerName="marketplace-operator" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663568 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerName="marketplace-operator" Sep 30 16:15:19 crc kubenswrapper[4796]: E0930 16:15:19.663588 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663597 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="extract-utilities" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663756 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77743d1-a6ef-4741-83d2-a102d2a256dc" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663777 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7881eaa3-77bc-430c-a2a6-d44ada95cc0a" containerName="marketplace-operator" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663793 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f6d193-b5f7-49cc-acf0-1698e5a15de7" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663809 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="78afec2a-ad1c-4999-8165-eda2ed82ef5d" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.663824 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0e5842f-b782-477d-87d7-c677e786c155" containerName="registry-server" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.664950 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.672570 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.674484 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8r5j"] Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.849274 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/399cf9ef-2274-4e40-bf07-f279f6d3730c-utilities\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.849893 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96qrj\" (UniqueName: \"kubernetes.io/projected/399cf9ef-2274-4e40-bf07-f279f6d3730c-kube-api-access-96qrj\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.850056 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/399cf9ef-2274-4e40-bf07-f279f6d3730c-catalog-content\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.862051 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n28l7"] Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.863186 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.867762 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.874170 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n28l7"] Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.951169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/399cf9ef-2274-4e40-bf07-f279f6d3730c-utilities\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.951213 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96qrj\" (UniqueName: \"kubernetes.io/projected/399cf9ef-2274-4e40-bf07-f279f6d3730c-kube-api-access-96qrj\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.951257 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/399cf9ef-2274-4e40-bf07-f279f6d3730c-catalog-content\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.952141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/399cf9ef-2274-4e40-bf07-f279f6d3730c-utilities\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.952172 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/399cf9ef-2274-4e40-bf07-f279f6d3730c-catalog-content\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.973273 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96qrj\" (UniqueName: \"kubernetes.io/projected/399cf9ef-2274-4e40-bf07-f279f6d3730c-kube-api-access-96qrj\") pod \"redhat-marketplace-n8r5j\" (UID: \"399cf9ef-2274-4e40-bf07-f279f6d3730c\") " pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:19 crc kubenswrapper[4796]: I0930 16:15:19.987545 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.053233 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wcpw\" (UniqueName: \"kubernetes.io/projected/227b1cd2-3c9b-44c2-baa9-e29119fe40de-kube-api-access-2wcpw\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.053349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/227b1cd2-3c9b-44c2-baa9-e29119fe40de-utilities\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.053410 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/227b1cd2-3c9b-44c2-baa9-e29119fe40de-catalog-content\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.154337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wcpw\" (UniqueName: \"kubernetes.io/projected/227b1cd2-3c9b-44c2-baa9-e29119fe40de-kube-api-access-2wcpw\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.154804 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/227b1cd2-3c9b-44c2-baa9-e29119fe40de-utilities\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.154833 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/227b1cd2-3c9b-44c2-baa9-e29119fe40de-catalog-content\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.155385 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/227b1cd2-3c9b-44c2-baa9-e29119fe40de-catalog-content\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.155502 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/227b1cd2-3c9b-44c2-baa9-e29119fe40de-utilities\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.177523 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wcpw\" (UniqueName: \"kubernetes.io/projected/227b1cd2-3c9b-44c2-baa9-e29119fe40de-kube-api-access-2wcpw\") pod \"redhat-operators-n28l7\" (UID: \"227b1cd2-3c9b-44c2-baa9-e29119fe40de\") " pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.188315 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8r5j"] Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.199389 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:20 crc kubenswrapper[4796]: W0930 16:15:20.200813 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod399cf9ef_2274_4e40_bf07_f279f6d3730c.slice/crio-a68d6d1fefc8b56d9783430f6cb9c69047fa1f62cede8aa3e7a787b199440cec WatchSource:0}: Error finding container a68d6d1fefc8b56d9783430f6cb9c69047fa1f62cede8aa3e7a787b199440cec: Status 404 returned error can't find the container with id a68d6d1fefc8b56d9783430f6cb9c69047fa1f62cede8aa3e7a787b199440cec Sep 30 16:15:20 crc kubenswrapper[4796]: I0930 16:15:20.402311 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n28l7"] Sep 30 16:15:21 crc kubenswrapper[4796]: I0930 16:15:21.142296 4796 generic.go:334] "Generic (PLEG): container finished" podID="227b1cd2-3c9b-44c2-baa9-e29119fe40de" containerID="f45f779b75ffc7f432361d42d924fc63ca2940dc78452f428c86c94e2f4648e5" exitCode=0 Sep 30 16:15:21 crc kubenswrapper[4796]: I0930 16:15:21.142385 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28l7" event={"ID":"227b1cd2-3c9b-44c2-baa9-e29119fe40de","Type":"ContainerDied","Data":"f45f779b75ffc7f432361d42d924fc63ca2940dc78452f428c86c94e2f4648e5"} Sep 30 16:15:21 crc kubenswrapper[4796]: I0930 16:15:21.142830 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28l7" event={"ID":"227b1cd2-3c9b-44c2-baa9-e29119fe40de","Type":"ContainerStarted","Data":"a249424dcdbf49cfa8be31421bb0da8dec2d5a9892baf3cbde5920e0e78b2f58"} Sep 30 16:15:21 crc kubenswrapper[4796]: I0930 16:15:21.144696 4796 generic.go:334] "Generic (PLEG): container finished" podID="399cf9ef-2274-4e40-bf07-f279f6d3730c" containerID="c9bc93486009caa7fe38d72e9f05bfa2ecfee4b78cef28b5b31e745ba7a8cebf" exitCode=0 Sep 30 16:15:21 crc kubenswrapper[4796]: I0930 16:15:21.145558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8r5j" event={"ID":"399cf9ef-2274-4e40-bf07-f279f6d3730c","Type":"ContainerDied","Data":"c9bc93486009caa7fe38d72e9f05bfa2ecfee4b78cef28b5b31e745ba7a8cebf"} Sep 30 16:15:21 crc kubenswrapper[4796]: I0930 16:15:21.145640 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8r5j" event={"ID":"399cf9ef-2274-4e40-bf07-f279f6d3730c","Type":"ContainerStarted","Data":"a68d6d1fefc8b56d9783430f6cb9c69047fa1f62cede8aa3e7a787b199440cec"} Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.060752 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zw4ct"] Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.062119 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.064174 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.070840 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw4ct"] Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.081814 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0723eea3-9e01-4945-be27-a9bbfd46cb3c-catalog-content\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.081854 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0723eea3-9e01-4945-be27-a9bbfd46cb3c-utilities\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.081897 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw89z\" (UniqueName: \"kubernetes.io/projected/0723eea3-9e01-4945-be27-a9bbfd46cb3c-kube-api-access-sw89z\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.182966 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw89z\" (UniqueName: \"kubernetes.io/projected/0723eea3-9e01-4945-be27-a9bbfd46cb3c-kube-api-access-sw89z\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.183877 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0723eea3-9e01-4945-be27-a9bbfd46cb3c-catalog-content\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.184049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0723eea3-9e01-4945-be27-a9bbfd46cb3c-utilities\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.184320 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0723eea3-9e01-4945-be27-a9bbfd46cb3c-catalog-content\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.184355 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0723eea3-9e01-4945-be27-a9bbfd46cb3c-utilities\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.206413 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw89z\" (UniqueName: \"kubernetes.io/projected/0723eea3-9e01-4945-be27-a9bbfd46cb3c-kube-api-access-sw89z\") pod \"certified-operators-zw4ct\" (UID: \"0723eea3-9e01-4945-be27-a9bbfd46cb3c\") " pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.261823 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gx5s8"] Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.263113 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.264865 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.272933 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gx5s8"] Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.284943 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5nbz\" (UniqueName: \"kubernetes.io/projected/cbe44b77-d340-4354-b289-f6adb481aa27-kube-api-access-t5nbz\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.285017 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-utilities\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.285058 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-catalog-content\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.380752 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.386424 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-utilities\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.386477 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-catalog-content\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.386729 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5nbz\" (UniqueName: \"kubernetes.io/projected/cbe44b77-d340-4354-b289-f6adb481aa27-kube-api-access-t5nbz\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.386850 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-utilities\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.386870 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-catalog-content\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.403567 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5nbz\" (UniqueName: \"kubernetes.io/projected/cbe44b77-d340-4354-b289-f6adb481aa27-kube-api-access-t5nbz\") pod \"community-operators-gx5s8\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.585445 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw4ct"] Sep 30 16:15:22 crc kubenswrapper[4796]: W0930 16:15:22.590934 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0723eea3_9e01_4945_be27_a9bbfd46cb3c.slice/crio-1ab9828e6173d8f6d8654747ba886409545a57b6218cc4f616bf0ab4cffb12f8 WatchSource:0}: Error finding container 1ab9828e6173d8f6d8654747ba886409545a57b6218cc4f616bf0ab4cffb12f8: Status 404 returned error can't find the container with id 1ab9828e6173d8f6d8654747ba886409545a57b6218cc4f616bf0ab4cffb12f8 Sep 30 16:15:22 crc kubenswrapper[4796]: I0930 16:15:22.684734 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.086149 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gx5s8"] Sep 30 16:15:23 crc kubenswrapper[4796]: W0930 16:15:23.092756 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbe44b77_d340_4354_b289_f6adb481aa27.slice/crio-d16410c31923bc0bc25bca57102353f0e3fc4db8655ff933a32af79b692c7c1c WatchSource:0}: Error finding container d16410c31923bc0bc25bca57102353f0e3fc4db8655ff933a32af79b692c7c1c: Status 404 returned error can't find the container with id d16410c31923bc0bc25bca57102353f0e3fc4db8655ff933a32af79b692c7c1c Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.158662 4796 generic.go:334] "Generic (PLEG): container finished" podID="0723eea3-9e01-4945-be27-a9bbfd46cb3c" containerID="932f20015b553ca862485f020b53a812380db5fcc7e6a5b09a099f4d68433402" exitCode=0 Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.158755 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4ct" event={"ID":"0723eea3-9e01-4945-be27-a9bbfd46cb3c","Type":"ContainerDied","Data":"932f20015b553ca862485f020b53a812380db5fcc7e6a5b09a099f4d68433402"} Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.158786 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4ct" event={"ID":"0723eea3-9e01-4945-be27-a9bbfd46cb3c","Type":"ContainerStarted","Data":"1ab9828e6173d8f6d8654747ba886409545a57b6218cc4f616bf0ab4cffb12f8"} Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.160968 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx5s8" event={"ID":"cbe44b77-d340-4354-b289-f6adb481aa27","Type":"ContainerStarted","Data":"d16410c31923bc0bc25bca57102353f0e3fc4db8655ff933a32af79b692c7c1c"} Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.167448 4796 generic.go:334] "Generic (PLEG): container finished" podID="399cf9ef-2274-4e40-bf07-f279f6d3730c" containerID="87044b0f221dcb29b3b6400bd0aeadfb1e6a6ad9ee4b2aa8e3e639163336e64a" exitCode=0 Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.167547 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8r5j" event={"ID":"399cf9ef-2274-4e40-bf07-f279f6d3730c","Type":"ContainerDied","Data":"87044b0f221dcb29b3b6400bd0aeadfb1e6a6ad9ee4b2aa8e3e639163336e64a"} Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.173301 4796 generic.go:334] "Generic (PLEG): container finished" podID="227b1cd2-3c9b-44c2-baa9-e29119fe40de" containerID="cc519cb4591428f85ce93f8d77870a253e005b9a3825dd11bd4a9d83d36c3a53" exitCode=0 Sep 30 16:15:23 crc kubenswrapper[4796]: I0930 16:15:23.173340 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28l7" event={"ID":"227b1cd2-3c9b-44c2-baa9-e29119fe40de","Type":"ContainerDied","Data":"cc519cb4591428f85ce93f8d77870a253e005b9a3825dd11bd4a9d83d36c3a53"} Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.186901 4796 generic.go:334] "Generic (PLEG): container finished" podID="cbe44b77-d340-4354-b289-f6adb481aa27" containerID="6de655918e681cbd1a0dae7542884eb1d3d332e1580d6f0e623c662a7a96e33a" exitCode=0 Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.187064 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx5s8" event={"ID":"cbe44b77-d340-4354-b289-f6adb481aa27","Type":"ContainerDied","Data":"6de655918e681cbd1a0dae7542884eb1d3d332e1580d6f0e623c662a7a96e33a"} Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.198119 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8r5j" event={"ID":"399cf9ef-2274-4e40-bf07-f279f6d3730c","Type":"ContainerStarted","Data":"36a7b1a54b882dd31056fbb7ae5d663644157164c15022d13e52c4ab0d70bb61"} Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.201854 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28l7" event={"ID":"227b1cd2-3c9b-44c2-baa9-e29119fe40de","Type":"ContainerStarted","Data":"ab4b49db3e1df467a0e879c21b0a7534e58299ff3a684365c53640b0d7d60266"} Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.206515 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4ct" event={"ID":"0723eea3-9e01-4945-be27-a9bbfd46cb3c","Type":"ContainerStarted","Data":"02cd92ede6a1baf35867dc5aeb4c2a76e6437843700b6ec9b4116bd33b5204f0"} Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.231420 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n28l7" podStartSLOduration=2.75165168 podStartE2EDuration="5.231404191s" podCreationTimestamp="2025-09-30 16:15:19 +0000 UTC" firstStartedPulling="2025-09-30 16:15:21.145140604 +0000 UTC m=+213.158419131" lastFinishedPulling="2025-09-30 16:15:23.624893115 +0000 UTC m=+215.638171642" observedRunningTime="2025-09-30 16:15:24.228507789 +0000 UTC m=+216.241786316" watchObservedRunningTime="2025-09-30 16:15:24.231404191 +0000 UTC m=+216.244682718" Sep 30 16:15:24 crc kubenswrapper[4796]: I0930 16:15:24.247474 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n8r5j" podStartSLOduration=2.778789791 podStartE2EDuration="5.247454997s" podCreationTimestamp="2025-09-30 16:15:19 +0000 UTC" firstStartedPulling="2025-09-30 16:15:21.146948525 +0000 UTC m=+213.160227052" lastFinishedPulling="2025-09-30 16:15:23.615613731 +0000 UTC m=+215.628892258" observedRunningTime="2025-09-30 16:15:24.24438016 +0000 UTC m=+216.257658687" watchObservedRunningTime="2025-09-30 16:15:24.247454997 +0000 UTC m=+216.260733524" Sep 30 16:15:25 crc kubenswrapper[4796]: I0930 16:15:25.216228 4796 generic.go:334] "Generic (PLEG): container finished" podID="0723eea3-9e01-4945-be27-a9bbfd46cb3c" containerID="02cd92ede6a1baf35867dc5aeb4c2a76e6437843700b6ec9b4116bd33b5204f0" exitCode=0 Sep 30 16:15:25 crc kubenswrapper[4796]: I0930 16:15:25.218222 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4ct" event={"ID":"0723eea3-9e01-4945-be27-a9bbfd46cb3c","Type":"ContainerDied","Data":"02cd92ede6a1baf35867dc5aeb4c2a76e6437843700b6ec9b4116bd33b5204f0"} Sep 30 16:15:26 crc kubenswrapper[4796]: I0930 16:15:26.223572 4796 generic.go:334] "Generic (PLEG): container finished" podID="cbe44b77-d340-4354-b289-f6adb481aa27" containerID="93456527f859d651ac332c836f05d1ffe6d42cc8bfe47102ff11d0ce76e5d272" exitCode=0 Sep 30 16:15:26 crc kubenswrapper[4796]: I0930 16:15:26.223620 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx5s8" event={"ID":"cbe44b77-d340-4354-b289-f6adb481aa27","Type":"ContainerDied","Data":"93456527f859d651ac332c836f05d1ffe6d42cc8bfe47102ff11d0ce76e5d272"} Sep 30 16:15:26 crc kubenswrapper[4796]: I0930 16:15:26.225900 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4ct" event={"ID":"0723eea3-9e01-4945-be27-a9bbfd46cb3c","Type":"ContainerStarted","Data":"c97da15688aa69128b7e34600084bf2cb6728edf559b3029af3fcf10b4077904"} Sep 30 16:15:26 crc kubenswrapper[4796]: I0930 16:15:26.265281 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zw4ct" podStartSLOduration=1.645283503 podStartE2EDuration="4.265264393s" podCreationTimestamp="2025-09-30 16:15:22 +0000 UTC" firstStartedPulling="2025-09-30 16:15:23.162686842 +0000 UTC m=+215.175965369" lastFinishedPulling="2025-09-30 16:15:25.782667732 +0000 UTC m=+217.795946259" observedRunningTime="2025-09-30 16:15:26.26128508 +0000 UTC m=+218.274563617" watchObservedRunningTime="2025-09-30 16:15:26.265264393 +0000 UTC m=+218.278542910" Sep 30 16:15:27 crc kubenswrapper[4796]: I0930 16:15:27.233089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx5s8" event={"ID":"cbe44b77-d340-4354-b289-f6adb481aa27","Type":"ContainerStarted","Data":"dc436f64dc0c06b0b89f0063900dd49204a4cbaafbaf5682372540e1696b08e3"} Sep 30 16:15:27 crc kubenswrapper[4796]: I0930 16:15:27.252807 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gx5s8" podStartSLOduration=2.698295387 podStartE2EDuration="5.252792264s" podCreationTimestamp="2025-09-30 16:15:22 +0000 UTC" firstStartedPulling="2025-09-30 16:15:24.190473837 +0000 UTC m=+216.203752364" lastFinishedPulling="2025-09-30 16:15:26.744970714 +0000 UTC m=+218.758249241" observedRunningTime="2025-09-30 16:15:27.251105936 +0000 UTC m=+219.264384513" watchObservedRunningTime="2025-09-30 16:15:27.252792264 +0000 UTC m=+219.266070791" Sep 30 16:15:29 crc kubenswrapper[4796]: I0930 16:15:29.987858 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:29 crc kubenswrapper[4796]: I0930 16:15:29.988448 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:30 crc kubenswrapper[4796]: I0930 16:15:30.044159 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:30 crc kubenswrapper[4796]: I0930 16:15:30.200426 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:30 crc kubenswrapper[4796]: I0930 16:15:30.200475 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:30 crc kubenswrapper[4796]: I0930 16:15:30.246307 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:30 crc kubenswrapper[4796]: I0930 16:15:30.287498 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n8r5j" Sep 30 16:15:30 crc kubenswrapper[4796]: I0930 16:15:30.310458 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n28l7" Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.096650 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.096724 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.096780 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.097324 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.097380 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc" gracePeriod=600 Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.260252 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc" exitCode=0 Sep 30 16:15:31 crc kubenswrapper[4796]: I0930 16:15:31.260330 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc"} Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.268213 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"5082179e6fe7c7d1bcb9bd9c5a623316782def0b2b5780c0f823267433a5c6e4"} Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.381283 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.381593 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.430358 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.685338 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.685527 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:32 crc kubenswrapper[4796]: I0930 16:15:32.740398 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:15:33 crc kubenswrapper[4796]: I0930 16:15:33.311083 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zw4ct" Sep 30 16:15:33 crc kubenswrapper[4796]: I0930 16:15:33.327269 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:17:31 crc kubenswrapper[4796]: I0930 16:17:31.096798 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:17:31 crc kubenswrapper[4796]: I0930 16:17:31.098190 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:18:01 crc kubenswrapper[4796]: I0930 16:18:01.097056 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:18:01 crc kubenswrapper[4796]: I0930 16:18:01.097935 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.097408 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.098346 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.098423 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.099323 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5082179e6fe7c7d1bcb9bd9c5a623316782def0b2b5780c0f823267433a5c6e4"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.099427 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://5082179e6fe7c7d1bcb9bd9c5a623316782def0b2b5780c0f823267433a5c6e4" gracePeriod=600 Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.446317 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="5082179e6fe7c7d1bcb9bd9c5a623316782def0b2b5780c0f823267433a5c6e4" exitCode=0 Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.446378 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"5082179e6fe7c7d1bcb9bd9c5a623316782def0b2b5780c0f823267433a5c6e4"} Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.446810 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"4ba0c7bae7193c789288a51223146b959225052eddbb4fa93010e8f66b95083d"} Sep 30 16:18:31 crc kubenswrapper[4796]: I0930 16:18:31.446843 4796 scope.go:117] "RemoveContainer" containerID="0da959fa7940f6e9db4df8973d531364ed58f26410653fe273eb783f1edb9acc" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.357805 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kdn56"] Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.359634 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.382129 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kdn56"] Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.541668 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cceb3071-9599-4110-b64f-18448ad2d35e-trusted-ca\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.541752 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cceb3071-9599-4110-b64f-18448ad2d35e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.541796 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-registry-tls\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.541844 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjqwk\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-kube-api-access-tjqwk\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.541886 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cceb3071-9599-4110-b64f-18448ad2d35e-registry-certificates\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.541919 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cceb3071-9599-4110-b64f-18448ad2d35e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.542154 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.542318 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-bound-sa-token\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.568529 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.643851 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-registry-tls\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.643953 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjqwk\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-kube-api-access-tjqwk\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.644027 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cceb3071-9599-4110-b64f-18448ad2d35e-registry-certificates\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.644066 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cceb3071-9599-4110-b64f-18448ad2d35e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.644142 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-bound-sa-token\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.644200 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cceb3071-9599-4110-b64f-18448ad2d35e-trusted-ca\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.644245 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cceb3071-9599-4110-b64f-18448ad2d35e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.645339 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cceb3071-9599-4110-b64f-18448ad2d35e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.646089 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cceb3071-9599-4110-b64f-18448ad2d35e-trusted-ca\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.646662 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cceb3071-9599-4110-b64f-18448ad2d35e-registry-certificates\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.656962 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-registry-tls\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.657052 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cceb3071-9599-4110-b64f-18448ad2d35e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.665081 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-bound-sa-token\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.671352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjqwk\" (UniqueName: \"kubernetes.io/projected/cceb3071-9599-4110-b64f-18448ad2d35e-kube-api-access-tjqwk\") pod \"image-registry-66df7c8f76-kdn56\" (UID: \"cceb3071-9599-4110-b64f-18448ad2d35e\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.683850 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:00 crc kubenswrapper[4796]: I0930 16:19:00.884382 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kdn56"] Sep 30 16:19:01 crc kubenswrapper[4796]: I0930 16:19:01.640873 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" event={"ID":"cceb3071-9599-4110-b64f-18448ad2d35e","Type":"ContainerStarted","Data":"5bbbb22ada6ef19a2687281ffb3b8db0d0facb25ffb7be2ec56f6ff036b41ad4"} Sep 30 16:19:01 crc kubenswrapper[4796]: I0930 16:19:01.640933 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" event={"ID":"cceb3071-9599-4110-b64f-18448ad2d35e","Type":"ContainerStarted","Data":"555f99d60c036f91e1586228f35da41b14a1b65283767f50ec346e19d475baa1"} Sep 30 16:19:01 crc kubenswrapper[4796]: I0930 16:19:01.641171 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:01 crc kubenswrapper[4796]: I0930 16:19:01.665683 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" podStartSLOduration=1.665652081 podStartE2EDuration="1.665652081s" podCreationTimestamp="2025-09-30 16:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:19:01.65750965 +0000 UTC m=+433.670788177" watchObservedRunningTime="2025-09-30 16:19:01.665652081 +0000 UTC m=+433.678930638" Sep 30 16:19:20 crc kubenswrapper[4796]: I0930 16:19:20.689130 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-kdn56" Sep 30 16:19:20 crc kubenswrapper[4796]: I0930 16:19:20.765025 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fl5fz"] Sep 30 16:19:45 crc kubenswrapper[4796]: I0930 16:19:45.819713 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" podUID="8aba7c6e-3d24-4217-9687-93a535675dbc" containerName="registry" containerID="cri-o://c2da43cb5e55daaa18fabf2be3caf53ed82144610d785157935b601e098ebe4b" gracePeriod=30 Sep 30 16:19:45 crc kubenswrapper[4796]: I0930 16:19:45.953220 4796 generic.go:334] "Generic (PLEG): container finished" podID="8aba7c6e-3d24-4217-9687-93a535675dbc" containerID="c2da43cb5e55daaa18fabf2be3caf53ed82144610d785157935b601e098ebe4b" exitCode=0 Sep 30 16:19:45 crc kubenswrapper[4796]: I0930 16:19:45.953411 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" event={"ID":"8aba7c6e-3d24-4217-9687-93a535675dbc","Type":"ContainerDied","Data":"c2da43cb5e55daaa18fabf2be3caf53ed82144610d785157935b601e098ebe4b"} Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.213531 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336113 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336228 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8aba7c6e-3d24-4217-9687-93a535675dbc-installation-pull-secrets\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336306 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-tls\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336381 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzt5s\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-kube-api-access-xzt5s\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336470 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-certificates\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336601 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8aba7c6e-3d24-4217-9687-93a535675dbc-ca-trust-extracted\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336649 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-bound-sa-token\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.336708 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-trusted-ca\") pod \"8aba7c6e-3d24-4217-9687-93a535675dbc\" (UID: \"8aba7c6e-3d24-4217-9687-93a535675dbc\") " Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.337905 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.338073 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.346381 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-kube-api-access-xzt5s" (OuterVolumeSpecName: "kube-api-access-xzt5s") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "kube-api-access-xzt5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.346589 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aba7c6e-3d24-4217-9687-93a535675dbc-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.347583 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.348107 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.356704 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.358170 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aba7c6e-3d24-4217-9687-93a535675dbc-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8aba7c6e-3d24-4217-9687-93a535675dbc" (UID: "8aba7c6e-3d24-4217-9687-93a535675dbc"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438334 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438393 4796 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8aba7c6e-3d24-4217-9687-93a535675dbc-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438414 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438433 4796 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8aba7c6e-3d24-4217-9687-93a535675dbc-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438457 4796 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438475 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzt5s\" (UniqueName: \"kubernetes.io/projected/8aba7c6e-3d24-4217-9687-93a535675dbc-kube-api-access-xzt5s\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.438491 4796 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8aba7c6e-3d24-4217-9687-93a535675dbc-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.961548 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" event={"ID":"8aba7c6e-3d24-4217-9687-93a535675dbc","Type":"ContainerDied","Data":"123e66d98bbb2b844640ae619e030b4ead9bf588bdaca5c91cc93601692cad91"} Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.961615 4796 scope.go:117] "RemoveContainer" containerID="c2da43cb5e55daaa18fabf2be3caf53ed82144610d785157935b601e098ebe4b" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.961627 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fl5fz" Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.983879 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fl5fz"] Sep 30 16:19:46 crc kubenswrapper[4796]: I0930 16:19:46.988586 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fl5fz"] Sep 30 16:19:48 crc kubenswrapper[4796]: I0930 16:19:48.747742 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aba7c6e-3d24-4217-9687-93a535675dbc" path="/var/lib/kubelet/pods/8aba7c6e-3d24-4217-9687-93a535675dbc/volumes" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.907212 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-zl4xg"] Sep 30 16:20:22 crc kubenswrapper[4796]: E0930 16:20:22.908339 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aba7c6e-3d24-4217-9687-93a535675dbc" containerName="registry" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.908357 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aba7c6e-3d24-4217-9687-93a535675dbc" containerName="registry" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.908481 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aba7c6e-3d24-4217-9687-93a535675dbc" containerName="registry" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.908942 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.914604 4796 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-c4sxl" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.914630 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.914962 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.915490 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-s2jmq"] Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.916259 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-s2jmq" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.918324 4796 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-nkc4q" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.922072 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-zl4xg"] Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.937930 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hgxtd"] Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.938652 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.941480 4796 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qcwws" Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.975634 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-s2jmq"] Sep 30 16:20:22 crc kubenswrapper[4796]: I0930 16:20:22.989902 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hgxtd"] Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.009137 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2pws\" (UniqueName: \"kubernetes.io/projected/4bd447e5-097a-4b02-be6e-6af942a45369-kube-api-access-r2pws\") pod \"cert-manager-cainjector-7f985d654d-zl4xg\" (UID: \"4bd447e5-097a-4b02-be6e-6af942a45369\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.009192 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csgrk\" (UniqueName: \"kubernetes.io/projected/df8942ab-8a15-4023-a58f-c36972391914-kube-api-access-csgrk\") pod \"cert-manager-webhook-5655c58dd6-hgxtd\" (UID: \"df8942ab-8a15-4023-a58f-c36972391914\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.009270 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs66r\" (UniqueName: \"kubernetes.io/projected/8245bb2a-a295-4493-b9fc-88469011a23c-kube-api-access-zs66r\") pod \"cert-manager-5b446d88c5-s2jmq\" (UID: \"8245bb2a-a295-4493-b9fc-88469011a23c\") " pod="cert-manager/cert-manager-5b446d88c5-s2jmq" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.110173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs66r\" (UniqueName: \"kubernetes.io/projected/8245bb2a-a295-4493-b9fc-88469011a23c-kube-api-access-zs66r\") pod \"cert-manager-5b446d88c5-s2jmq\" (UID: \"8245bb2a-a295-4493-b9fc-88469011a23c\") " pod="cert-manager/cert-manager-5b446d88c5-s2jmq" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.110257 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2pws\" (UniqueName: \"kubernetes.io/projected/4bd447e5-097a-4b02-be6e-6af942a45369-kube-api-access-r2pws\") pod \"cert-manager-cainjector-7f985d654d-zl4xg\" (UID: \"4bd447e5-097a-4b02-be6e-6af942a45369\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.110295 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csgrk\" (UniqueName: \"kubernetes.io/projected/df8942ab-8a15-4023-a58f-c36972391914-kube-api-access-csgrk\") pod \"cert-manager-webhook-5655c58dd6-hgxtd\" (UID: \"df8942ab-8a15-4023-a58f-c36972391914\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.131728 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2pws\" (UniqueName: \"kubernetes.io/projected/4bd447e5-097a-4b02-be6e-6af942a45369-kube-api-access-r2pws\") pod \"cert-manager-cainjector-7f985d654d-zl4xg\" (UID: \"4bd447e5-097a-4b02-be6e-6af942a45369\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.131728 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs66r\" (UniqueName: \"kubernetes.io/projected/8245bb2a-a295-4493-b9fc-88469011a23c-kube-api-access-zs66r\") pod \"cert-manager-5b446d88c5-s2jmq\" (UID: \"8245bb2a-a295-4493-b9fc-88469011a23c\") " pod="cert-manager/cert-manager-5b446d88c5-s2jmq" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.132323 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csgrk\" (UniqueName: \"kubernetes.io/projected/df8942ab-8a15-4023-a58f-c36972391914-kube-api-access-csgrk\") pod \"cert-manager-webhook-5655c58dd6-hgxtd\" (UID: \"df8942ab-8a15-4023-a58f-c36972391914\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.227169 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.235313 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-s2jmq" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.254308 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.474553 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-zl4xg"] Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.488520 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.496439 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-s2jmq"] Sep 30 16:20:23 crc kubenswrapper[4796]: W0930 16:20:23.503427 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8245bb2a_a295_4493_b9fc_88469011a23c.slice/crio-5103b2d7fa1b8b28544b1159ea9efb1bdd87c69316f00ef208b64c0e8866f348 WatchSource:0}: Error finding container 5103b2d7fa1b8b28544b1159ea9efb1bdd87c69316f00ef208b64c0e8866f348: Status 404 returned error can't find the container with id 5103b2d7fa1b8b28544b1159ea9efb1bdd87c69316f00ef208b64c0e8866f348 Sep 30 16:20:23 crc kubenswrapper[4796]: I0930 16:20:23.531485 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hgxtd"] Sep 30 16:20:23 crc kubenswrapper[4796]: W0930 16:20:23.535031 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf8942ab_8a15_4023_a58f_c36972391914.slice/crio-b509a6848e11ec52e3e335dc89178514eab8ec7f528fbe8c15918b66fcef03d8 WatchSource:0}: Error finding container b509a6848e11ec52e3e335dc89178514eab8ec7f528fbe8c15918b66fcef03d8: Status 404 returned error can't find the container with id b509a6848e11ec52e3e335dc89178514eab8ec7f528fbe8c15918b66fcef03d8 Sep 30 16:20:24 crc kubenswrapper[4796]: I0930 16:20:24.199012 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" event={"ID":"df8942ab-8a15-4023-a58f-c36972391914","Type":"ContainerStarted","Data":"b509a6848e11ec52e3e335dc89178514eab8ec7f528fbe8c15918b66fcef03d8"} Sep 30 16:20:24 crc kubenswrapper[4796]: I0930 16:20:24.210569 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" event={"ID":"4bd447e5-097a-4b02-be6e-6af942a45369","Type":"ContainerStarted","Data":"f0982eaa5e8754f7d5e709bae6e78724e0f041274fe07683b91e539422e93d24"} Sep 30 16:20:24 crc kubenswrapper[4796]: I0930 16:20:24.215659 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-s2jmq" event={"ID":"8245bb2a-a295-4493-b9fc-88469011a23c","Type":"ContainerStarted","Data":"5103b2d7fa1b8b28544b1159ea9efb1bdd87c69316f00ef208b64c0e8866f348"} Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.233306 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" event={"ID":"4bd447e5-097a-4b02-be6e-6af942a45369","Type":"ContainerStarted","Data":"a64a5885e9fe17b643bfed906b2178689693b52fbaf2b562562010ca98bb633d"} Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.235137 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-s2jmq" event={"ID":"8245bb2a-a295-4493-b9fc-88469011a23c","Type":"ContainerStarted","Data":"489a68923904efba666993f5d54cb0120f35d9d1c9b2ae14986b1a97a0b179f8"} Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.238867 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" event={"ID":"df8942ab-8a15-4023-a58f-c36972391914","Type":"ContainerStarted","Data":"d30e2afef822c874a535d6957c617930c1edbfd67c98eab2bd3ea3d16a16f63c"} Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.239546 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.253850 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-zl4xg" podStartSLOduration=1.8851563420000002 podStartE2EDuration="5.253827798s" podCreationTimestamp="2025-09-30 16:20:22 +0000 UTC" firstStartedPulling="2025-09-30 16:20:23.488288122 +0000 UTC m=+515.501566649" lastFinishedPulling="2025-09-30 16:20:26.856959568 +0000 UTC m=+518.870238105" observedRunningTime="2025-09-30 16:20:27.253388625 +0000 UTC m=+519.266667162" watchObservedRunningTime="2025-09-30 16:20:27.253827798 +0000 UTC m=+519.267106325" Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.273379 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" podStartSLOduration=1.879211899 podStartE2EDuration="5.27334934s" podCreationTimestamp="2025-09-30 16:20:22 +0000 UTC" firstStartedPulling="2025-09-30 16:20:23.536835248 +0000 UTC m=+515.550113775" lastFinishedPulling="2025-09-30 16:20:26.930972649 +0000 UTC m=+518.944251216" observedRunningTime="2025-09-30 16:20:27.270246789 +0000 UTC m=+519.283525346" watchObservedRunningTime="2025-09-30 16:20:27.27334934 +0000 UTC m=+519.286627887" Sep 30 16:20:27 crc kubenswrapper[4796]: I0930 16:20:27.285971 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-s2jmq" podStartSLOduration=1.934131056 podStartE2EDuration="5.28594527s" podCreationTimestamp="2025-09-30 16:20:22 +0000 UTC" firstStartedPulling="2025-09-30 16:20:23.506154843 +0000 UTC m=+515.519433370" lastFinishedPulling="2025-09-30 16:20:26.857969057 +0000 UTC m=+518.871247584" observedRunningTime="2025-09-30 16:20:27.284804066 +0000 UTC m=+519.298082593" watchObservedRunningTime="2025-09-30 16:20:27.28594527 +0000 UTC m=+519.299223807" Sep 30 16:20:31 crc kubenswrapper[4796]: I0930 16:20:31.096921 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:20:31 crc kubenswrapper[4796]: I0930 16:20:31.097546 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.258693 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-hgxtd" Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.595249 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c6n4h"] Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.595876 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-controller" containerID="cri-o://5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.596229 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="sbdb" containerID="cri-o://cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.596346 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.596408 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-acl-logging" containerID="cri-o://074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.596255 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-node" containerID="cri-o://f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.596612 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="nbdb" containerID="cri-o://dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.599445 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="northd" containerID="cri-o://27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.636271 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" containerID="cri-o://dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" gracePeriod=30 Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.941086 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/2.log" Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.944089 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovn-acl-logging/0.log" Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.944756 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovn-controller/0.log" Sep 30 16:20:33 crc kubenswrapper[4796]: I0930 16:20:33.945203 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015657 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-g2z6w"] Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015845 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-node" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015858 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-node" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015867 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015873 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015884 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-acl-logging" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015889 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-acl-logging" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015899 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015905 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015912 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="northd" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015917 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="northd" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015926 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015931 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015939 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="nbdb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015944 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="nbdb" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015951 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015956 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.015965 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="sbdb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.015971 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="sbdb" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.016002 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kubecfg-setup" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016008 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kubecfg-setup" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.016020 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016026 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016110 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="sbdb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016122 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="northd" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016131 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-acl-logging" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016137 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovn-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016145 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016152 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="nbdb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016159 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016166 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-node" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016173 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016180 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.016279 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016287 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.016365 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerName="ovnkube-controller" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.017781 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066090 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-script-lib\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066133 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-kubelet\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066156 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-slash\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066176 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-openvswitch\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066241 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-slash" (OuterVolumeSpecName: "host-slash") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066241 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066341 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066376 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-config\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066402 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-env-overrides\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066444 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-node-log\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066470 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-etc-openvswitch\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066545 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-node-log" (OuterVolumeSpecName: "node-log") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066640 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066838 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066891 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-bin\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066893 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066913 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-var-lib-openvswitch\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.066961 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067023 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-netns\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067035 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067047 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-ovn-kubernetes\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067069 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067078 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7df6e69-ff3a-484a-b440-8afe12028b40-ovn-node-metrics-cert\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067106 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067137 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-systemd-units\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067127 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067159 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmcms\" (UniqueName: \"kubernetes.io/projected/f7df6e69-ff3a-484a-b440-8afe12028b40-kube-api-access-jmcms\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067194 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067213 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067274 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-ovn\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067341 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-systemd\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067371 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-log-socket\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067400 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-netd\") pod \"f7df6e69-ff3a-484a-b440-8afe12028b40\" (UID: \"f7df6e69-ff3a-484a-b440-8afe12028b40\") " Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067820 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067835 4796 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067846 4796 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-node-log\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067859 4796 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067869 4796 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067879 4796 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067890 4796 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067901 4796 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067911 4796 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067922 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7df6e69-ff3a-484a-b440-8afe12028b40-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067933 4796 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067942 4796 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-slash\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067952 4796 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.067314 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.068007 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.068047 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-log-socket" (OuterVolumeSpecName: "log-socket") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.068064 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.075399 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7df6e69-ff3a-484a-b440-8afe12028b40-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.075528 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7df6e69-ff3a-484a-b440-8afe12028b40-kube-api-access-jmcms" (OuterVolumeSpecName: "kube-api-access-jmcms") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "kube-api-access-jmcms". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.083178 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f7df6e69-ff3a-484a-b440-8afe12028b40" (UID: "f7df6e69-ff3a-484a-b440-8afe12028b40"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169251 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-run-ovn-kubernetes\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169306 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-log-socket\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169481 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-slash\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169620 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-systemd-units\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169695 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c199057d-06e1-427c-9945-d1ed71b229a0-ovn-node-metrics-cert\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169760 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169813 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-ovnkube-config\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169891 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-cni-netd\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.169960 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-run-netns\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170105 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-var-lib-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170167 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-systemd\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170259 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d79xk\" (UniqueName: \"kubernetes.io/projected/c199057d-06e1-427c-9945-d1ed71b229a0-kube-api-access-d79xk\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170331 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-ovn\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170421 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-env-overrides\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170477 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170559 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-cni-bin\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170727 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-etc-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170815 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-kubelet\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170851 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-node-log\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.170876 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-ovnkube-script-lib\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171040 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7df6e69-ff3a-484a-b440-8afe12028b40-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171061 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmcms\" (UniqueName: \"kubernetes.io/projected/f7df6e69-ff3a-484a-b440-8afe12028b40-kube-api-access-jmcms\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171119 4796 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171162 4796 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171204 4796 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-log-socket\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171229 4796 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.171262 4796 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7df6e69-ff3a-484a-b440-8afe12028b40-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272143 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-slash\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-systemd-units\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272223 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c199057d-06e1-427c-9945-d1ed71b229a0-ovn-node-metrics-cert\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272244 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272263 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-ovnkube-config\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272282 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-cni-netd\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272298 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-run-netns\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272318 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-var-lib-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-systemd\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272359 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d79xk\" (UniqueName: \"kubernetes.io/projected/c199057d-06e1-427c-9945-d1ed71b229a0-kube-api-access-d79xk\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272385 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-ovn\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272398 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-systemd-units\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272410 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-env-overrides\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272465 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272498 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-cni-bin\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272531 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-etc-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272527 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-var-lib-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272562 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-kubelet\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272583 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-node-log\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-ovnkube-script-lib\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272626 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-run-ovn-kubernetes\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272630 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-cni-netd\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-log-socket\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272650 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272707 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-log-socket\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272714 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-systemd\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272678 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-run-netns\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272754 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-etc-openvswitch\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272770 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-cni-bin\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272825 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-run-ovn\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272840 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-node-log\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272870 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-slash\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272888 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-kubelet\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272925 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-run-ovn-kubernetes\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.272458 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c199057d-06e1-427c-9945-d1ed71b229a0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.273164 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-env-overrides\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.273454 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-ovnkube-config\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.273742 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c199057d-06e1-427c-9945-d1ed71b229a0-ovnkube-script-lib\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.277379 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c199057d-06e1-427c-9945-d1ed71b229a0-ovn-node-metrics-cert\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.288719 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hldp6_b4cf8ea3-7aa4-4dfc-8443-347fd3898f15/kube-multus/1.log" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.289420 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hldp6_b4cf8ea3-7aa4-4dfc-8443-347fd3898f15/kube-multus/0.log" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.289520 4796 generic.go:334] "Generic (PLEG): container finished" podID="b4cf8ea3-7aa4-4dfc-8443-347fd3898f15" containerID="24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc" exitCode=2 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.289575 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hldp6" event={"ID":"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15","Type":"ContainerDied","Data":"24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.289655 4796 scope.go:117] "RemoveContainer" containerID="1fb523a7b08aa5c49a7d4ec2ce7b333495d40c67f7047a8bb3a4a2412bf0409f" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.290606 4796 scope.go:117] "RemoveContainer" containerID="24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.291905 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-hldp6_openshift-multus(b4cf8ea3-7aa4-4dfc-8443-347fd3898f15)\"" pod="openshift-multus/multus-hldp6" podUID="b4cf8ea3-7aa4-4dfc-8443-347fd3898f15" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.292575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d79xk\" (UniqueName: \"kubernetes.io/projected/c199057d-06e1-427c-9945-d1ed71b229a0-kube-api-access-d79xk\") pod \"ovnkube-node-g2z6w\" (UID: \"c199057d-06e1-427c-9945-d1ed71b229a0\") " pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.295285 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovnkube-controller/2.log" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.298159 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovn-acl-logging/0.log" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.298950 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c6n4h_f7df6e69-ff3a-484a-b440-8afe12028b40/ovn-controller/0.log" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299458 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" exitCode=0 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299491 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" exitCode=0 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299505 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" exitCode=0 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299521 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" exitCode=0 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299534 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" exitCode=0 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299535 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299592 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299612 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299547 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" exitCode=0 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299656 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299670 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" exitCode=143 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299838 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7df6e69-ff3a-484a-b440-8afe12028b40" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" exitCode=143 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299652 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299876 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299894 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299911 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299922 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299932 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299942 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299952 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299961 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.299972 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300028 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300040 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300054 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300072 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300084 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300096 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300106 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300117 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300126 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300136 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300149 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300159 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300169 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300184 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300199 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300211 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300221 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300230 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300239 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300249 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300258 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300268 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300278 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300288 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300302 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6n4h" event={"ID":"f7df6e69-ff3a-484a-b440-8afe12028b40","Type":"ContainerDied","Data":"7f6acb784913fb54b9caa46e850bbf8d7467eabc4a7306815d95e9a2951b8925"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300319 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300332 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300343 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300354 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300364 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300374 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300384 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300393 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300403 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.300413 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.331461 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.343297 4796 scope.go:117] "RemoveContainer" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.343429 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c6n4h"] Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.353173 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c6n4h"] Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.364956 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:20:34 crc kubenswrapper[4796]: W0930 16:20:34.377211 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc199057d_06e1_427c_9945_d1ed71b229a0.slice/crio-d44396659b8f023502ca8884ff187fc7d69e5e9d0a23f2a53b3474343eb66ca6 WatchSource:0}: Error finding container d44396659b8f023502ca8884ff187fc7d69e5e9d0a23f2a53b3474343eb66ca6: Status 404 returned error can't find the container with id d44396659b8f023502ca8884ff187fc7d69e5e9d0a23f2a53b3474343eb66ca6 Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.392227 4796 scope.go:117] "RemoveContainer" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.411935 4796 scope.go:117] "RemoveContainer" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.425206 4796 scope.go:117] "RemoveContainer" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.445893 4796 scope.go:117] "RemoveContainer" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.464608 4796 scope.go:117] "RemoveContainer" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.479392 4796 scope.go:117] "RemoveContainer" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.491375 4796 scope.go:117] "RemoveContainer" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.577891 4796 scope.go:117] "RemoveContainer" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.612857 4796 scope.go:117] "RemoveContainer" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.613284 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": container with ID starting with dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282 not found: ID does not exist" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.613308 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} err="failed to get container status \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": rpc error: code = NotFound desc = could not find container \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": container with ID starting with dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.613330 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.613857 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": container with ID starting with f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808 not found: ID does not exist" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.613883 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} err="failed to get container status \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": rpc error: code = NotFound desc = could not find container \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": container with ID starting with f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.613902 4796 scope.go:117] "RemoveContainer" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.614167 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": container with ID starting with cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28 not found: ID does not exist" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.614190 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} err="failed to get container status \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": rpc error: code = NotFound desc = could not find container \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": container with ID starting with cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.614204 4796 scope.go:117] "RemoveContainer" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.614527 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": container with ID starting with dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3 not found: ID does not exist" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.614556 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} err="failed to get container status \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": rpc error: code = NotFound desc = could not find container \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": container with ID starting with dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.614570 4796 scope.go:117] "RemoveContainer" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.614801 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": container with ID starting with 27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37 not found: ID does not exist" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.614830 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} err="failed to get container status \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": rpc error: code = NotFound desc = could not find container \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": container with ID starting with 27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.614848 4796 scope.go:117] "RemoveContainer" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.615284 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": container with ID starting with 20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107 not found: ID does not exist" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.615312 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} err="failed to get container status \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": rpc error: code = NotFound desc = could not find container \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": container with ID starting with 20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.615325 4796 scope.go:117] "RemoveContainer" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.615552 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": container with ID starting with f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979 not found: ID does not exist" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.615579 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} err="failed to get container status \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": rpc error: code = NotFound desc = could not find container \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": container with ID starting with f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.615594 4796 scope.go:117] "RemoveContainer" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.615864 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": container with ID starting with 074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb not found: ID does not exist" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.615884 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} err="failed to get container status \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": rpc error: code = NotFound desc = could not find container \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": container with ID starting with 074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.615898 4796 scope.go:117] "RemoveContainer" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.616256 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": container with ID starting with 5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e not found: ID does not exist" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.616287 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} err="failed to get container status \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": rpc error: code = NotFound desc = could not find container \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": container with ID starting with 5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.616303 4796 scope.go:117] "RemoveContainer" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" Sep 30 16:20:34 crc kubenswrapper[4796]: E0930 16:20:34.616558 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": container with ID starting with ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e not found: ID does not exist" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.616584 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} err="failed to get container status \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": rpc error: code = NotFound desc = could not find container \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": container with ID starting with ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.616601 4796 scope.go:117] "RemoveContainer" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.619273 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} err="failed to get container status \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": rpc error: code = NotFound desc = could not find container \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": container with ID starting with dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.619299 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.619765 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} err="failed to get container status \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": rpc error: code = NotFound desc = could not find container \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": container with ID starting with f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.619852 4796 scope.go:117] "RemoveContainer" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.621016 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} err="failed to get container status \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": rpc error: code = NotFound desc = could not find container \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": container with ID starting with cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.621042 4796 scope.go:117] "RemoveContainer" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.621392 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} err="failed to get container status \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": rpc error: code = NotFound desc = could not find container \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": container with ID starting with dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.621432 4796 scope.go:117] "RemoveContainer" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.621930 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} err="failed to get container status \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": rpc error: code = NotFound desc = could not find container \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": container with ID starting with 27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.621958 4796 scope.go:117] "RemoveContainer" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.622378 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} err="failed to get container status \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": rpc error: code = NotFound desc = could not find container \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": container with ID starting with 20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.622428 4796 scope.go:117] "RemoveContainer" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.622770 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} err="failed to get container status \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": rpc error: code = NotFound desc = could not find container \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": container with ID starting with f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.622823 4796 scope.go:117] "RemoveContainer" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.623127 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} err="failed to get container status \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": rpc error: code = NotFound desc = could not find container \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": container with ID starting with 074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.623152 4796 scope.go:117] "RemoveContainer" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.623578 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} err="failed to get container status \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": rpc error: code = NotFound desc = could not find container \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": container with ID starting with 5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.623615 4796 scope.go:117] "RemoveContainer" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.623928 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} err="failed to get container status \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": rpc error: code = NotFound desc = could not find container \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": container with ID starting with ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.623974 4796 scope.go:117] "RemoveContainer" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.624349 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} err="failed to get container status \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": rpc error: code = NotFound desc = could not find container \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": container with ID starting with dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.624374 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.624670 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} err="failed to get container status \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": rpc error: code = NotFound desc = could not find container \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": container with ID starting with f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.624697 4796 scope.go:117] "RemoveContainer" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.625018 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} err="failed to get container status \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": rpc error: code = NotFound desc = could not find container \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": container with ID starting with cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.625038 4796 scope.go:117] "RemoveContainer" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.625302 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} err="failed to get container status \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": rpc error: code = NotFound desc = could not find container \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": container with ID starting with dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.625339 4796 scope.go:117] "RemoveContainer" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.625650 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} err="failed to get container status \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": rpc error: code = NotFound desc = could not find container \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": container with ID starting with 27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.625685 4796 scope.go:117] "RemoveContainer" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.626005 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} err="failed to get container status \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": rpc error: code = NotFound desc = could not find container \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": container with ID starting with 20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.626031 4796 scope.go:117] "RemoveContainer" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.626349 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} err="failed to get container status \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": rpc error: code = NotFound desc = could not find container \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": container with ID starting with f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.626370 4796 scope.go:117] "RemoveContainer" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.626821 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} err="failed to get container status \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": rpc error: code = NotFound desc = could not find container \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": container with ID starting with 074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.626901 4796 scope.go:117] "RemoveContainer" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.627374 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} err="failed to get container status \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": rpc error: code = NotFound desc = could not find container \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": container with ID starting with 5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.627401 4796 scope.go:117] "RemoveContainer" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.627802 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} err="failed to get container status \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": rpc error: code = NotFound desc = could not find container \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": container with ID starting with ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.627965 4796 scope.go:117] "RemoveContainer" containerID="dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.628507 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282"} err="failed to get container status \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": rpc error: code = NotFound desc = could not find container \"dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282\": container with ID starting with dbc12f552444c3df322a7e4559558a3fef879fde04c0636dba1a40ee45f96282 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.628550 4796 scope.go:117] "RemoveContainer" containerID="f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.628830 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808"} err="failed to get container status \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": rpc error: code = NotFound desc = could not find container \"f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808\": container with ID starting with f92f7ab37e8f90b2d12e870b7920caa64eb5571c714ec86bf55347aab41aa808 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.628880 4796 scope.go:117] "RemoveContainer" containerID="cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.629228 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28"} err="failed to get container status \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": rpc error: code = NotFound desc = could not find container \"cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28\": container with ID starting with cfbd7ea9ebf258cb1080197fe3bc3e2ae0a6f6f5d50eaaf14b29fa0305a91e28 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.629258 4796 scope.go:117] "RemoveContainer" containerID="dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.629803 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3"} err="failed to get container status \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": rpc error: code = NotFound desc = could not find container \"dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3\": container with ID starting with dcba428ca07b4131551b31df8e258267796b6a8b18abd12577c373ed915ddcf3 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.629825 4796 scope.go:117] "RemoveContainer" containerID="27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.630190 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37"} err="failed to get container status \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": rpc error: code = NotFound desc = could not find container \"27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37\": container with ID starting with 27c661c5914a70fc324aebc45356a8ee38ca2c246563639d8194cc6bee40bd37 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.630242 4796 scope.go:117] "RemoveContainer" containerID="20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.630762 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107"} err="failed to get container status \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": rpc error: code = NotFound desc = could not find container \"20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107\": container with ID starting with 20cd40d9419c1d6557087dd11fb779137930c4b4c9ec17b98dd5b524cbbac107 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.630793 4796 scope.go:117] "RemoveContainer" containerID="f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.631162 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979"} err="failed to get container status \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": rpc error: code = NotFound desc = could not find container \"f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979\": container with ID starting with f055da4f17f65c6aee9684a47c58a84680c76bf9a11719cd42a8b507f0268979 not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.631225 4796 scope.go:117] "RemoveContainer" containerID="074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.631668 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb"} err="failed to get container status \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": rpc error: code = NotFound desc = could not find container \"074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb\": container with ID starting with 074388b670e182ee3c20d7014a2684499f3c63f6f38b458e8d570e7a57f01cfb not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.631693 4796 scope.go:117] "RemoveContainer" containerID="5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.632184 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e"} err="failed to get container status \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": rpc error: code = NotFound desc = could not find container \"5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e\": container with ID starting with 5a9e9b0c9f7abe2de88f7cc815a5c3691f9fee78bc94dfc19295ff493552fe8e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.632210 4796 scope.go:117] "RemoveContainer" containerID="ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.632578 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e"} err="failed to get container status \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": rpc error: code = NotFound desc = could not find container \"ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e\": container with ID starting with ceed638f654cabb845e0b5583c216220d81af18582e9cd5870323d5f16d9812e not found: ID does not exist" Sep 30 16:20:34 crc kubenswrapper[4796]: I0930 16:20:34.740224 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7df6e69-ff3a-484a-b440-8afe12028b40" path="/var/lib/kubelet/pods/f7df6e69-ff3a-484a-b440-8afe12028b40/volumes" Sep 30 16:20:35 crc kubenswrapper[4796]: I0930 16:20:35.309494 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hldp6_b4cf8ea3-7aa4-4dfc-8443-347fd3898f15/kube-multus/1.log" Sep 30 16:20:35 crc kubenswrapper[4796]: I0930 16:20:35.313952 4796 generic.go:334] "Generic (PLEG): container finished" podID="c199057d-06e1-427c-9945-d1ed71b229a0" containerID="9bbf9f0ba2fc91b0e68cf516585a670585b4be913b400585ae985785b7a3286a" exitCode=0 Sep 30 16:20:35 crc kubenswrapper[4796]: I0930 16:20:35.314054 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerDied","Data":"9bbf9f0ba2fc91b0e68cf516585a670585b4be913b400585ae985785b7a3286a"} Sep 30 16:20:35 crc kubenswrapper[4796]: I0930 16:20:35.314153 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"d44396659b8f023502ca8884ff187fc7d69e5e9d0a23f2a53b3474343eb66ca6"} Sep 30 16:20:36 crc kubenswrapper[4796]: I0930 16:20:36.327459 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"8426132362c49be234aa62cc79690ece101f9a2c047c203f8167bcb68328b3a0"} Sep 30 16:20:36 crc kubenswrapper[4796]: I0930 16:20:36.327929 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"ab6ff48905a272261dbdb6a6b06e908205ad4ccfc19e1f0c23df2eb0e527fa01"} Sep 30 16:20:36 crc kubenswrapper[4796]: I0930 16:20:36.327953 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"843d99c097d93ba13a56adcde0c16eb3f51b585ce159d15fc6406ee98637ab8b"} Sep 30 16:20:36 crc kubenswrapper[4796]: I0930 16:20:36.327973 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"48da1d57916d6564b992f27c70dd4c907cfe638eec2f6c05c7b602ff93a66d85"} Sep 30 16:20:36 crc kubenswrapper[4796]: I0930 16:20:36.328021 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"60bab3db8554c52f256cf0f1615915138694b6f2b217a995bbb705ddcd194c6d"} Sep 30 16:20:36 crc kubenswrapper[4796]: I0930 16:20:36.328038 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"c1e94d6ba94db972be3b579f135e58db2892fcd0f3118201149107e7ba95f960"} Sep 30 16:20:39 crc kubenswrapper[4796]: I0930 16:20:39.358757 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"8994a53ec6baf4bf1b9f0875435093924d47127ac4a462e043dc14cead8dec60"} Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.374941 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" event={"ID":"c199057d-06e1-427c-9945-d1ed71b229a0","Type":"ContainerStarted","Data":"b52d394246246a623ee7dda9717b8dca212a56ebd263cd4e1cfbd979be3b7832"} Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.375455 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.375473 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.375486 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.405826 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" podStartSLOduration=8.405807058 podStartE2EDuration="8.405807058s" podCreationTimestamp="2025-09-30 16:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:20:41.403992755 +0000 UTC m=+533.417271282" watchObservedRunningTime="2025-09-30 16:20:41.405807058 +0000 UTC m=+533.419085585" Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.409646 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:41 crc kubenswrapper[4796]: I0930 16:20:41.413717 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:20:45 crc kubenswrapper[4796]: I0930 16:20:45.732772 4796 scope.go:117] "RemoveContainer" containerID="24b820a752e343ec060d42f2d745e6b33618df5e16d6092b3a8b997df451dcbc" Sep 30 16:20:46 crc kubenswrapper[4796]: I0930 16:20:46.417377 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hldp6_b4cf8ea3-7aa4-4dfc-8443-347fd3898f15/kube-multus/1.log" Sep 30 16:20:46 crc kubenswrapper[4796]: I0930 16:20:46.417959 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hldp6" event={"ID":"b4cf8ea3-7aa4-4dfc-8443-347fd3898f15","Type":"ContainerStarted","Data":"dc1696382f79c4bf3b9f6fa6ddb502d2b709ea1fb3912c3196fad2e41551610b"} Sep 30 16:20:57 crc kubenswrapper[4796]: I0930 16:20:57.988371 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj"] Sep 30 16:20:57 crc kubenswrapper[4796]: I0930 16:20:57.990661 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:57 crc kubenswrapper[4796]: I0930 16:20:57.994174 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.009761 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj"] Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.015026 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qs2p\" (UniqueName: \"kubernetes.io/projected/c5876a04-724a-4dda-a51d-d84db778caf8-kube-api-access-5qs2p\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.015134 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-bundle\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.015203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-util\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.116609 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-util\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.116791 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qs2p\" (UniqueName: \"kubernetes.io/projected/c5876a04-724a-4dda-a51d-d84db778caf8-kube-api-access-5qs2p\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.116878 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-bundle\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.117325 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-util\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.117573 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-bundle\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.158494 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qs2p\" (UniqueName: \"kubernetes.io/projected/c5876a04-724a-4dda-a51d-d84db778caf8-kube-api-access-5qs2p\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.182005 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7"] Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.183283 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.195474 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7"] Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.218294 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-util\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.218341 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-bundle\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.218363 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5gcv\" (UniqueName: \"kubernetes.io/projected/f910fbb3-44e8-4db1-a65b-5fb0297676bf-kube-api-access-q5gcv\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.313353 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.319609 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-bundle\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.319672 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5gcv\" (UniqueName: \"kubernetes.io/projected/f910fbb3-44e8-4db1-a65b-5fb0297676bf-kube-api-access-q5gcv\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.319773 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-util\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.320316 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-bundle\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.320396 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-util\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.342709 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5gcv\" (UniqueName: \"kubernetes.io/projected/f910fbb3-44e8-4db1-a65b-5fb0297676bf-kube-api-access-q5gcv\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.508859 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.513242 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj"] Sep 30 16:20:58 crc kubenswrapper[4796]: W0930 16:20:58.530971 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5876a04_724a_4dda_a51d_d84db778caf8.slice/crio-308974b702ef1fbaaf6a69b70a5b1cbe9b8713a03f8f68dd4b2b4082741ff7b6 WatchSource:0}: Error finding container 308974b702ef1fbaaf6a69b70a5b1cbe9b8713a03f8f68dd4b2b4082741ff7b6: Status 404 returned error can't find the container with id 308974b702ef1fbaaf6a69b70a5b1cbe9b8713a03f8f68dd4b2b4082741ff7b6 Sep 30 16:20:58 crc kubenswrapper[4796]: I0930 16:20:58.765558 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7"] Sep 30 16:20:58 crc kubenswrapper[4796]: W0930 16:20:58.766834 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf910fbb3_44e8_4db1_a65b_5fb0297676bf.slice/crio-2d7b70bb0042eaea0c1b6f11890adee7514dd644772a912d83a1731bae3801c1 WatchSource:0}: Error finding container 2d7b70bb0042eaea0c1b6f11890adee7514dd644772a912d83a1731bae3801c1: Status 404 returned error can't find the container with id 2d7b70bb0042eaea0c1b6f11890adee7514dd644772a912d83a1731bae3801c1 Sep 30 16:20:59 crc kubenswrapper[4796]: I0930 16:20:59.508020 4796 generic.go:334] "Generic (PLEG): container finished" podID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerID="d6cdbe8a3ef6ef74e7ac7b38e48bd283512490e6d945b295398bb5b77967f4ce" exitCode=0 Sep 30 16:20:59 crc kubenswrapper[4796]: I0930 16:20:59.508123 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" event={"ID":"f910fbb3-44e8-4db1-a65b-5fb0297676bf","Type":"ContainerDied","Data":"d6cdbe8a3ef6ef74e7ac7b38e48bd283512490e6d945b295398bb5b77967f4ce"} Sep 30 16:20:59 crc kubenswrapper[4796]: I0930 16:20:59.508168 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" event={"ID":"f910fbb3-44e8-4db1-a65b-5fb0297676bf","Type":"ContainerStarted","Data":"2d7b70bb0042eaea0c1b6f11890adee7514dd644772a912d83a1731bae3801c1"} Sep 30 16:20:59 crc kubenswrapper[4796]: I0930 16:20:59.510463 4796 generic.go:334] "Generic (PLEG): container finished" podID="c5876a04-724a-4dda-a51d-d84db778caf8" containerID="372aed981cef931e53fdbf998d7921fa2efeb096bc3d23031f442ad0bd070397" exitCode=0 Sep 30 16:20:59 crc kubenswrapper[4796]: I0930 16:20:59.510530 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" event={"ID":"c5876a04-724a-4dda-a51d-d84db778caf8","Type":"ContainerDied","Data":"372aed981cef931e53fdbf998d7921fa2efeb096bc3d23031f442ad0bd070397"} Sep 30 16:20:59 crc kubenswrapper[4796]: I0930 16:20:59.510569 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" event={"ID":"c5876a04-724a-4dda-a51d-d84db778caf8","Type":"ContainerStarted","Data":"308974b702ef1fbaaf6a69b70a5b1cbe9b8713a03f8f68dd4b2b4082741ff7b6"} Sep 30 16:21:01 crc kubenswrapper[4796]: I0930 16:21:01.097614 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:21:01 crc kubenswrapper[4796]: I0930 16:21:01.098168 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:21:01 crc kubenswrapper[4796]: I0930 16:21:01.528074 4796 generic.go:334] "Generic (PLEG): container finished" podID="c5876a04-724a-4dda-a51d-d84db778caf8" containerID="fbfae3f9eecca7a2d2856a412252cd90cea0762862695078c6ecfd7f87fb9456" exitCode=0 Sep 30 16:21:01 crc kubenswrapper[4796]: I0930 16:21:01.528145 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" event={"ID":"c5876a04-724a-4dda-a51d-d84db778caf8","Type":"ContainerDied","Data":"fbfae3f9eecca7a2d2856a412252cd90cea0762862695078c6ecfd7f87fb9456"} Sep 30 16:21:01 crc kubenswrapper[4796]: I0930 16:21:01.532202 4796 generic.go:334] "Generic (PLEG): container finished" podID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerID="fba43e29f5bafd0b35c5eb5fde8056931c700c0f81c8f68827bd60045a62a100" exitCode=0 Sep 30 16:21:01 crc kubenswrapper[4796]: I0930 16:21:01.532364 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" event={"ID":"f910fbb3-44e8-4db1-a65b-5fb0297676bf","Type":"ContainerDied","Data":"fba43e29f5bafd0b35c5eb5fde8056931c700c0f81c8f68827bd60045a62a100"} Sep 30 16:21:02 crc kubenswrapper[4796]: I0930 16:21:02.541761 4796 generic.go:334] "Generic (PLEG): container finished" podID="c5876a04-724a-4dda-a51d-d84db778caf8" containerID="8ff3769beb0bb6f427d96b19cd5cd3696cb14beae758d5afd1b1c414089dea03" exitCode=0 Sep 30 16:21:02 crc kubenswrapper[4796]: I0930 16:21:02.541857 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" event={"ID":"c5876a04-724a-4dda-a51d-d84db778caf8","Type":"ContainerDied","Data":"8ff3769beb0bb6f427d96b19cd5cd3696cb14beae758d5afd1b1c414089dea03"} Sep 30 16:21:02 crc kubenswrapper[4796]: I0930 16:21:02.545384 4796 generic.go:334] "Generic (PLEG): container finished" podID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerID="39b5ddd86ed383d0c57ce05c4349ebbf0ce8eb23140318bdd345feb37d49d75f" exitCode=0 Sep 30 16:21:02 crc kubenswrapper[4796]: I0930 16:21:02.545438 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" event={"ID":"f910fbb3-44e8-4db1-a65b-5fb0297676bf","Type":"ContainerDied","Data":"39b5ddd86ed383d0c57ce05c4349ebbf0ce8eb23140318bdd345feb37d49d75f"} Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.887476 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.893006 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.906582 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5gcv\" (UniqueName: \"kubernetes.io/projected/f910fbb3-44e8-4db1-a65b-5fb0297676bf-kube-api-access-q5gcv\") pod \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.906646 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-bundle\") pod \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.906709 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qs2p\" (UniqueName: \"kubernetes.io/projected/c5876a04-724a-4dda-a51d-d84db778caf8-kube-api-access-5qs2p\") pod \"c5876a04-724a-4dda-a51d-d84db778caf8\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.906733 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-util\") pod \"c5876a04-724a-4dda-a51d-d84db778caf8\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.906771 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-util\") pod \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\" (UID: \"f910fbb3-44e8-4db1-a65b-5fb0297676bf\") " Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.906827 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-bundle\") pod \"c5876a04-724a-4dda-a51d-d84db778caf8\" (UID: \"c5876a04-724a-4dda-a51d-d84db778caf8\") " Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.908199 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-bundle" (OuterVolumeSpecName: "bundle") pod "f910fbb3-44e8-4db1-a65b-5fb0297676bf" (UID: "f910fbb3-44e8-4db1-a65b-5fb0297676bf"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.911815 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-bundle" (OuterVolumeSpecName: "bundle") pod "c5876a04-724a-4dda-a51d-d84db778caf8" (UID: "c5876a04-724a-4dda-a51d-d84db778caf8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.915172 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5876a04-724a-4dda-a51d-d84db778caf8-kube-api-access-5qs2p" (OuterVolumeSpecName: "kube-api-access-5qs2p") pod "c5876a04-724a-4dda-a51d-d84db778caf8" (UID: "c5876a04-724a-4dda-a51d-d84db778caf8"). InnerVolumeSpecName "kube-api-access-5qs2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.916267 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f910fbb3-44e8-4db1-a65b-5fb0297676bf-kube-api-access-q5gcv" (OuterVolumeSpecName: "kube-api-access-q5gcv") pod "f910fbb3-44e8-4db1-a65b-5fb0297676bf" (UID: "f910fbb3-44e8-4db1-a65b-5fb0297676bf"). InnerVolumeSpecName "kube-api-access-q5gcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:21:03 crc kubenswrapper[4796]: I0930 16:21:03.921693 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-util" (OuterVolumeSpecName: "util") pod "f910fbb3-44e8-4db1-a65b-5fb0297676bf" (UID: "f910fbb3-44e8-4db1-a65b-5fb0297676bf"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.008902 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.009136 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.009186 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5gcv\" (UniqueName: \"kubernetes.io/projected/f910fbb3-44e8-4db1-a65b-5fb0297676bf-kube-api-access-q5gcv\") on node \"crc\" DevicePath \"\"" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.009212 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f910fbb3-44e8-4db1-a65b-5fb0297676bf-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.009222 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qs2p\" (UniqueName: \"kubernetes.io/projected/c5876a04-724a-4dda-a51d-d84db778caf8-kube-api-access-5qs2p\") on node \"crc\" DevicePath \"\"" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.319699 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-util" (OuterVolumeSpecName: "util") pod "c5876a04-724a-4dda-a51d-d84db778caf8" (UID: "c5876a04-724a-4dda-a51d-d84db778caf8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.375146 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g2z6w" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.415961 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5876a04-724a-4dda-a51d-d84db778caf8-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.560734 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.560753 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7" event={"ID":"f910fbb3-44e8-4db1-a65b-5fb0297676bf","Type":"ContainerDied","Data":"2d7b70bb0042eaea0c1b6f11890adee7514dd644772a912d83a1731bae3801c1"} Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.561274 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d7b70bb0042eaea0c1b6f11890adee7514dd644772a912d83a1731bae3801c1" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.562948 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" event={"ID":"c5876a04-724a-4dda-a51d-d84db778caf8","Type":"ContainerDied","Data":"308974b702ef1fbaaf6a69b70a5b1cbe9b8713a03f8f68dd4b2b4082741ff7b6"} Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.563010 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="308974b702ef1fbaaf6a69b70a5b1cbe9b8713a03f8f68dd4b2b4082741ff7b6" Sep 30 16:21:04 crc kubenswrapper[4796]: I0930 16:21:04.563126 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.546913 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-fcc886d58-q8fnc"] Sep 30 16:21:10 crc kubenswrapper[4796]: E0930 16:21:10.547560 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="pull" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547578 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="pull" Sep 30 16:21:10 crc kubenswrapper[4796]: E0930 16:21:10.547597 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="extract" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547603 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="extract" Sep 30 16:21:10 crc kubenswrapper[4796]: E0930 16:21:10.547615 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="util" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547621 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="util" Sep 30 16:21:10 crc kubenswrapper[4796]: E0930 16:21:10.547649 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="util" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547655 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="util" Sep 30 16:21:10 crc kubenswrapper[4796]: E0930 16:21:10.547668 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="extract" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547676 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="extract" Sep 30 16:21:10 crc kubenswrapper[4796]: E0930 16:21:10.547691 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="pull" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547696 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="pull" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547800 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f910fbb3-44e8-4db1-a65b-5fb0297676bf" containerName="extract" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.547818 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5876a04-724a-4dda-a51d-d84db778caf8" containerName="extract" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.548365 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.550600 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-xctss" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.551552 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.552247 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.558540 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-fcc886d58-q8fnc"] Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.597068 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcfdh\" (UniqueName: \"kubernetes.io/projected/ae187c99-f8dd-4c13-a001-3cb453b62739-kube-api-access-gcfdh\") pod \"cluster-logging-operator-fcc886d58-q8fnc\" (UID: \"ae187c99-f8dd-4c13-a001-3cb453b62739\") " pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.699195 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcfdh\" (UniqueName: \"kubernetes.io/projected/ae187c99-f8dd-4c13-a001-3cb453b62739-kube-api-access-gcfdh\") pod \"cluster-logging-operator-fcc886d58-q8fnc\" (UID: \"ae187c99-f8dd-4c13-a001-3cb453b62739\") " pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.734263 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcfdh\" (UniqueName: \"kubernetes.io/projected/ae187c99-f8dd-4c13-a001-3cb453b62739-kube-api-access-gcfdh\") pod \"cluster-logging-operator-fcc886d58-q8fnc\" (UID: \"ae187c99-f8dd-4c13-a001-3cb453b62739\") " pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" Sep 30 16:21:10 crc kubenswrapper[4796]: I0930 16:21:10.918373 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" Sep 30 16:21:11 crc kubenswrapper[4796]: I0930 16:21:11.140830 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-fcc886d58-q8fnc"] Sep 30 16:21:11 crc kubenswrapper[4796]: I0930 16:21:11.610075 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" event={"ID":"ae187c99-f8dd-4c13-a001-3cb453b62739","Type":"ContainerStarted","Data":"2d3ec2eb1584629331967848921b88081735e6793ebbfbbef07e979b222c569c"} Sep 30 16:21:17 crc kubenswrapper[4796]: I0930 16:21:17.661037 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" event={"ID":"ae187c99-f8dd-4c13-a001-3cb453b62739","Type":"ContainerStarted","Data":"6c4d21a9b9cbc5fae7e1efd7e02fa0d47e76c4aa0fc27f019463e6f177966696"} Sep 30 16:21:17 crc kubenswrapper[4796]: I0930 16:21:17.686410 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-fcc886d58-q8fnc" podStartSLOduration=2.2314723770000002 podStartE2EDuration="7.686391833s" podCreationTimestamp="2025-09-30 16:21:10 +0000 UTC" firstStartedPulling="2025-09-30 16:21:11.155208752 +0000 UTC m=+563.168487279" lastFinishedPulling="2025-09-30 16:21:16.610128208 +0000 UTC m=+568.623406735" observedRunningTime="2025-09-30 16:21:17.683929571 +0000 UTC m=+569.697208098" watchObservedRunningTime="2025-09-30 16:21:17.686391833 +0000 UTC m=+569.699670350" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.272365 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf"] Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.274396 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.291296 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.291790 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.292254 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-k49lb" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.297040 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.297056 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.298206 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.318631 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf"] Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.342749 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-webhook-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.342887 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/233c52ec-0b90-41d5-9442-bb0ce913adaf-manager-config\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.342927 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j497f\" (UniqueName: \"kubernetes.io/projected/233c52ec-0b90-41d5-9442-bb0ce913adaf-kube-api-access-j497f\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.342964 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.343015 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-apiservice-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.443809 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.443874 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-apiservice-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.443900 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-webhook-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.443956 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/233c52ec-0b90-41d5-9442-bb0ce913adaf-manager-config\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.443975 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j497f\" (UniqueName: \"kubernetes.io/projected/233c52ec-0b90-41d5-9442-bb0ce913adaf-kube-api-access-j497f\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.444917 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/233c52ec-0b90-41d5-9442-bb0ce913adaf-manager-config\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.450708 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-apiservice-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.451179 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.451612 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/233c52ec-0b90-41d5-9442-bb0ce913adaf-webhook-cert\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.462364 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j497f\" (UniqueName: \"kubernetes.io/projected/233c52ec-0b90-41d5-9442-bb0ce913adaf-kube-api-access-j497f\") pod \"loki-operator-controller-manager-769bbf7ffd-fzdwf\" (UID: \"233c52ec-0b90-41d5-9442-bb0ce913adaf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.591917 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:19 crc kubenswrapper[4796]: I0930 16:21:19.837968 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf"] Sep 30 16:21:20 crc kubenswrapper[4796]: I0930 16:21:20.681302 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" event={"ID":"233c52ec-0b90-41d5-9442-bb0ce913adaf","Type":"ContainerStarted","Data":"048d8a58efcfed1e658b7a5dc1998a3e6bbd6335c151332d12ecfc0998d0bb78"} Sep 30 16:21:23 crc kubenswrapper[4796]: I0930 16:21:23.708870 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" event={"ID":"233c52ec-0b90-41d5-9442-bb0ce913adaf","Type":"ContainerStarted","Data":"1811076d757e2297ccf9a6842d7c040bb5ade65ae9d9d2faed2a0fe0f7350b8a"} Sep 30 16:21:29 crc kubenswrapper[4796]: I0930 16:21:29.748036 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" event={"ID":"233c52ec-0b90-41d5-9442-bb0ce913adaf","Type":"ContainerStarted","Data":"b72ae31c5f341200fbdfac53145dd39da21fe27fff7f7196ace20e7090bb29cf"} Sep 30 16:21:29 crc kubenswrapper[4796]: I0930 16:21:29.750132 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:29 crc kubenswrapper[4796]: I0930 16:21:29.752428 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" Sep 30 16:21:29 crc kubenswrapper[4796]: I0930 16:21:29.800136 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-769bbf7ffd-fzdwf" podStartSLOduration=1.281820344 podStartE2EDuration="10.800120554s" podCreationTimestamp="2025-09-30 16:21:19 +0000 UTC" firstStartedPulling="2025-09-30 16:21:19.854239984 +0000 UTC m=+571.867518511" lastFinishedPulling="2025-09-30 16:21:29.372540194 +0000 UTC m=+581.385818721" observedRunningTime="2025-09-30 16:21:29.793351036 +0000 UTC m=+581.806629563" watchObservedRunningTime="2025-09-30 16:21:29.800120554 +0000 UTC m=+581.813399081" Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.097599 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.097668 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.097746 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.098406 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ba0c7bae7193c789288a51223146b959225052eddbb4fa93010e8f66b95083d"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.098462 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://4ba0c7bae7193c789288a51223146b959225052eddbb4fa93010e8f66b95083d" gracePeriod=600 Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.764100 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="4ba0c7bae7193c789288a51223146b959225052eddbb4fa93010e8f66b95083d" exitCode=0 Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.765904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"4ba0c7bae7193c789288a51223146b959225052eddbb4fa93010e8f66b95083d"} Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.765956 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"8e292f0f61906ba72a5d11ef43cd6aabb03ee9b745e2fef2028fac6c22e38f4f"} Sep 30 16:21:31 crc kubenswrapper[4796]: I0930 16:21:31.766004 4796 scope.go:117] "RemoveContainer" containerID="5082179e6fe7c7d1bcb9bd9c5a623316782def0b2b5780c0f823267433a5c6e4" Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.939354 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.940637 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.943071 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.943370 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.944254 4796 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-q7jgh" Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.954400 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.990902 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9djg\" (UniqueName: \"kubernetes.io/projected/ac9e0e23-974f-4d6c-9136-1a3be9770dda-kube-api-access-d9djg\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") " pod="minio-dev/minio" Sep 30 16:21:36 crc kubenswrapper[4796]: I0930 16:21:36.990949 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") " pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.091850 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9djg\" (UniqueName: \"kubernetes.io/projected/ac9e0e23-974f-4d6c-9136-1a3be9770dda-kube-api-access-d9djg\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") " pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.091897 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") " pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.094810 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.094860 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4c337e573384472f0f63ac79902baee65f02a6af03943354abccd76e6bbfd876/globalmount\"" pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.116949 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9djg\" (UniqueName: \"kubernetes.io/projected/ac9e0e23-974f-4d6c-9136-1a3be9770dda-kube-api-access-d9djg\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") " pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.134281 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-567b9570-b7fb-4cdf-9f2d-4d7ed0edac1d\") pod \"minio\" (UID: \"ac9e0e23-974f-4d6c-9136-1a3be9770dda\") " pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.260926 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.480229 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Sep 30 16:21:37 crc kubenswrapper[4796]: I0930 16:21:37.809910 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"ac9e0e23-974f-4d6c-9136-1a3be9770dda","Type":"ContainerStarted","Data":"7d2e358f2dbd4c48d9e75d129a1347368f2d18235c2bdf2b49c84553b53a7c0a"} Sep 30 16:21:41 crc kubenswrapper[4796]: I0930 16:21:41.836615 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"ac9e0e23-974f-4d6c-9136-1a3be9770dda","Type":"ContainerStarted","Data":"bbaca160afc899c93961442d503fc597cd4b79370acf0c62d3c192bcc5d6788c"} Sep 30 16:21:41 crc kubenswrapper[4796]: I0930 16:21:41.858269 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.292507063 podStartE2EDuration="7.858241321s" podCreationTimestamp="2025-09-30 16:21:34 +0000 UTC" firstStartedPulling="2025-09-30 16:21:37.493651455 +0000 UTC m=+589.506929982" lastFinishedPulling="2025-09-30 16:21:41.059385713 +0000 UTC m=+593.072664240" observedRunningTime="2025-09-30 16:21:41.854582014 +0000 UTC m=+593.867860601" watchObservedRunningTime="2025-09-30 16:21:41.858241321 +0000 UTC m=+593.871519848" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.057433 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-l7td6"] Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.059091 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.064029 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.064259 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-xk4kv" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.064425 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.064438 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.065462 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.079730 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.081340 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-l7td6"] Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093784 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da982abe-554b-4964-ae74-ec0b4fce349d-tmp\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093825 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-syslog-receiver\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093856 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-token\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093897 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-metrics\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093916 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-trusted-ca\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093932 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/da982abe-554b-4964-ae74-ec0b4fce349d-datadir\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.093989 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl8cf\" (UniqueName: \"kubernetes.io/projected/da982abe-554b-4964-ae74-ec0b4fce349d-kube-api-access-gl8cf\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.094007 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-config\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.094025 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-config-openshift-service-cacrt\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.094042 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/da982abe-554b-4964-ae74-ec0b4fce349d-sa-token\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.094069 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-entrypoint\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195367 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-config\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195422 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl8cf\" (UniqueName: \"kubernetes.io/projected/da982abe-554b-4964-ae74-ec0b4fce349d-kube-api-access-gl8cf\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195452 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-config-openshift-service-cacrt\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195484 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/da982abe-554b-4964-ae74-ec0b4fce349d-sa-token\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195517 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-entrypoint\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195568 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da982abe-554b-4964-ae74-ec0b4fce349d-tmp\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195589 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-syslog-receiver\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195617 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-token\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195662 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-metrics\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195683 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-trusted-ca\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195705 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/da982abe-554b-4964-ae74-ec0b4fce349d-datadir\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.195792 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/da982abe-554b-4964-ae74-ec0b4fce349d-datadir\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.196923 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-config\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: E0930 16:21:49.197360 4796 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Sep 30 16:21:49 crc kubenswrapper[4796]: E0930 16:21:49.197458 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-syslog-receiver podName:da982abe-554b-4964-ae74-ec0b4fce349d nodeName:}" failed. No retries permitted until 2025-09-30 16:21:49.69743894 +0000 UTC m=+601.710717467 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-syslog-receiver") pod "collector-l7td6" (UID: "da982abe-554b-4964-ae74-ec0b4fce349d") : secret "collector-syslog-receiver" not found Sep 30 16:21:49 crc kubenswrapper[4796]: E0930 16:21:49.197360 4796 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Sep 30 16:21:49 crc kubenswrapper[4796]: E0930 16:21:49.197568 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-metrics podName:da982abe-554b-4964-ae74-ec0b4fce349d nodeName:}" failed. No retries permitted until 2025-09-30 16:21:49.697539153 +0000 UTC m=+601.710817680 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-metrics") pod "collector-l7td6" (UID: "da982abe-554b-4964-ae74-ec0b4fce349d") : secret "collector-metrics" not found Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.197850 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-config-openshift-service-cacrt\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.198297 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-entrypoint\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.198615 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da982abe-554b-4964-ae74-ec0b4fce349d-trusted-ca\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.206317 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da982abe-554b-4964-ae74-ec0b4fce349d-tmp\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.207658 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-token\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.234546 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/da982abe-554b-4964-ae74-ec0b4fce349d-sa-token\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.239784 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl8cf\" (UniqueName: \"kubernetes.io/projected/da982abe-554b-4964-ae74-ec0b4fce349d-kube-api-access-gl8cf\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.703385 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-syslog-receiver\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.703874 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-metrics\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.708953 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-collector-syslog-receiver\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.710171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/da982abe-554b-4964-ae74-ec0b4fce349d-metrics\") pod \"collector-l7td6\" (UID: \"da982abe-554b-4964-ae74-ec0b4fce349d\") " pod="openshift-logging/collector-l7td6" Sep 30 16:21:49 crc kubenswrapper[4796]: I0930 16:21:49.987187 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-l7td6" Sep 30 16:21:50 crc kubenswrapper[4796]: I0930 16:21:50.256005 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-l7td6"] Sep 30 16:21:50 crc kubenswrapper[4796]: I0930 16:21:50.891677 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-l7td6" event={"ID":"da982abe-554b-4964-ae74-ec0b4fce349d","Type":"ContainerStarted","Data":"a9dd15a09b3f39ee26d6293d4709c4e677e25a7ae0f14aa997008141e7d3dffa"} Sep 30 16:21:58 crc kubenswrapper[4796]: I0930 16:21:58.938144 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-l7td6" event={"ID":"da982abe-554b-4964-ae74-ec0b4fce349d","Type":"ContainerStarted","Data":"beb9769baa76bacc2daa4f3c5587a5b193aa99abccc2ff05fd3ee1bde61edeaa"} Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.361230 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-l7td6" podStartSLOduration=10.829728969 podStartE2EDuration="18.36120631s" podCreationTimestamp="2025-09-30 16:21:49 +0000 UTC" firstStartedPulling="2025-09-30 16:21:50.259312796 +0000 UTC m=+602.272591333" lastFinishedPulling="2025-09-30 16:21:57.790790147 +0000 UTC m=+609.804068674" observedRunningTime="2025-09-30 16:21:58.965657488 +0000 UTC m=+610.978936055" watchObservedRunningTime="2025-09-30 16:22:07.36120631 +0000 UTC m=+619.374484837" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.362855 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2"] Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.364392 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.367035 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.377181 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2"] Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.487347 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.487403 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.487474 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bz67\" (UniqueName: \"kubernetes.io/projected/9a961850-d9b5-4e08-bbc6-cc2302da4a45-kube-api-access-6bz67\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.588524 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bz67\" (UniqueName: \"kubernetes.io/projected/9a961850-d9b5-4e08-bbc6-cc2302da4a45-kube-api-access-6bz67\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.588599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.588620 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.589304 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.589375 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.611131 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bz67\" (UniqueName: \"kubernetes.io/projected/9a961850-d9b5-4e08-bbc6-cc2302da4a45-kube-api-access-6bz67\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.686939 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.924167 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2"] Sep 30 16:22:07 crc kubenswrapper[4796]: I0930 16:22:07.990970 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" event={"ID":"9a961850-d9b5-4e08-bbc6-cc2302da4a45","Type":"ContainerStarted","Data":"f41e537fb7511d35010bbd09ffd5c2a0515ba0aad7e5112ed5758dd59b55ae34"} Sep 30 16:22:08 crc kubenswrapper[4796]: I0930 16:22:08.999687 4796 generic.go:334] "Generic (PLEG): container finished" podID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerID="9e0e05f3fed1f71d2731c2b23348d3639b4ecdee6b1a5465b860c3b61fdfa28b" exitCode=0 Sep 30 16:22:08 crc kubenswrapper[4796]: I0930 16:22:08.999740 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" event={"ID":"9a961850-d9b5-4e08-bbc6-cc2302da4a45","Type":"ContainerDied","Data":"9e0e05f3fed1f71d2731c2b23348d3639b4ecdee6b1a5465b860c3b61fdfa28b"} Sep 30 16:22:11 crc kubenswrapper[4796]: I0930 16:22:11.016177 4796 generic.go:334] "Generic (PLEG): container finished" podID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerID="69a71772e844e7efcd759c8296044c565b6c41d5e0e15182f3c3e3b161b98e3c" exitCode=0 Sep 30 16:22:11 crc kubenswrapper[4796]: I0930 16:22:11.016233 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" event={"ID":"9a961850-d9b5-4e08-bbc6-cc2302da4a45","Type":"ContainerDied","Data":"69a71772e844e7efcd759c8296044c565b6c41d5e0e15182f3c3e3b161b98e3c"} Sep 30 16:22:12 crc kubenswrapper[4796]: I0930 16:22:12.025168 4796 generic.go:334] "Generic (PLEG): container finished" podID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerID="da41f85cbf043a4f631e00b7d9df7de31be1744471248d58ddfa2e4c1d911422" exitCode=0 Sep 30 16:22:12 crc kubenswrapper[4796]: I0930 16:22:12.025220 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" event={"ID":"9a961850-d9b5-4e08-bbc6-cc2302da4a45","Type":"ContainerDied","Data":"da41f85cbf043a4f631e00b7d9df7de31be1744471248d58ddfa2e4c1d911422"} Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.305575 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.391368 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-bundle\") pod \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.391435 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-util\") pod \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.391522 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bz67\" (UniqueName: \"kubernetes.io/projected/9a961850-d9b5-4e08-bbc6-cc2302da4a45-kube-api-access-6bz67\") pod \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\" (UID: \"9a961850-d9b5-4e08-bbc6-cc2302da4a45\") " Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.392257 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-bundle" (OuterVolumeSpecName: "bundle") pod "9a961850-d9b5-4e08-bbc6-cc2302da4a45" (UID: "9a961850-d9b5-4e08-bbc6-cc2302da4a45"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.398196 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a961850-d9b5-4e08-bbc6-cc2302da4a45-kube-api-access-6bz67" (OuterVolumeSpecName: "kube-api-access-6bz67") pod "9a961850-d9b5-4e08-bbc6-cc2302da4a45" (UID: "9a961850-d9b5-4e08-bbc6-cc2302da4a45"). InnerVolumeSpecName "kube-api-access-6bz67". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.404622 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-util" (OuterVolumeSpecName: "util") pod "9a961850-d9b5-4e08-bbc6-cc2302da4a45" (UID: "9a961850-d9b5-4e08-bbc6-cc2302da4a45"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.492484 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bz67\" (UniqueName: \"kubernetes.io/projected/9a961850-d9b5-4e08-bbc6-cc2302da4a45-kube-api-access-6bz67\") on node \"crc\" DevicePath \"\"" Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.492538 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:22:13 crc kubenswrapper[4796]: I0930 16:22:13.492551 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a961850-d9b5-4e08-bbc6-cc2302da4a45-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:22:14 crc kubenswrapper[4796]: I0930 16:22:14.041723 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" event={"ID":"9a961850-d9b5-4e08-bbc6-cc2302da4a45","Type":"ContainerDied","Data":"f41e537fb7511d35010bbd09ffd5c2a0515ba0aad7e5112ed5758dd59b55ae34"} Sep 30 16:22:14 crc kubenswrapper[4796]: I0930 16:22:14.042226 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f41e537fb7511d35010bbd09ffd5c2a0515ba0aad7e5112ed5758dd59b55ae34" Sep 30 16:22:14 crc kubenswrapper[4796]: I0930 16:22:14.042265 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.134233 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs"] Sep 30 16:22:17 crc kubenswrapper[4796]: E0930 16:22:17.134931 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="util" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.134947 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="util" Sep 30 16:22:17 crc kubenswrapper[4796]: E0930 16:22:17.134960 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="pull" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.134967 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="pull" Sep 30 16:22:17 crc kubenswrapper[4796]: E0930 16:22:17.135008 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="extract" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.135017 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="extract" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.135129 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a961850-d9b5-4e08-bbc6-cc2302da4a45" containerName="extract" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.135574 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" Sep 30 16:22:17 crc kubenswrapper[4796]: W0930 16:22:17.138137 4796 reflector.go:561] object-"openshift-nmstate"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Sep 30 16:22:17 crc kubenswrapper[4796]: E0930 16:22:17.138311 4796 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.139632 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dvrdr" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.144293 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.147966 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs"] Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.239365 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gdxx\" (UniqueName: \"kubernetes.io/projected/70baf97d-1c3a-4106-9754-9369d381faf1-kube-api-access-6gdxx\") pod \"nmstate-operator-5d6f6cfd66-jpgzs\" (UID: \"70baf97d-1c3a-4106-9754-9369d381faf1\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" Sep 30 16:22:17 crc kubenswrapper[4796]: I0930 16:22:17.340548 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gdxx\" (UniqueName: \"kubernetes.io/projected/70baf97d-1c3a-4106-9754-9369d381faf1-kube-api-access-6gdxx\") pod \"nmstate-operator-5d6f6cfd66-jpgzs\" (UID: \"70baf97d-1c3a-4106-9754-9369d381faf1\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" Sep 30 16:22:18 crc kubenswrapper[4796]: E0930 16:22:18.352805 4796 projected.go:288] Couldn't get configMap openshift-nmstate/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Sep 30 16:22:18 crc kubenswrapper[4796]: E0930 16:22:18.352862 4796 projected.go:194] Error preparing data for projected volume kube-api-access-6gdxx for pod openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs: failed to sync configmap cache: timed out waiting for the condition Sep 30 16:22:18 crc kubenswrapper[4796]: E0930 16:22:18.352931 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70baf97d-1c3a-4106-9754-9369d381faf1-kube-api-access-6gdxx podName:70baf97d-1c3a-4106-9754-9369d381faf1 nodeName:}" failed. No retries permitted until 2025-09-30 16:22:18.852908553 +0000 UTC m=+630.866187100 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6gdxx" (UniqueName: "kubernetes.io/projected/70baf97d-1c3a-4106-9754-9369d381faf1-kube-api-access-6gdxx") pod "nmstate-operator-5d6f6cfd66-jpgzs" (UID: "70baf97d-1c3a-4106-9754-9369d381faf1") : failed to sync configmap cache: timed out waiting for the condition Sep 30 16:22:18 crc kubenswrapper[4796]: I0930 16:22:18.392811 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 30 16:22:18 crc kubenswrapper[4796]: I0930 16:22:18.860018 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gdxx\" (UniqueName: \"kubernetes.io/projected/70baf97d-1c3a-4106-9754-9369d381faf1-kube-api-access-6gdxx\") pod \"nmstate-operator-5d6f6cfd66-jpgzs\" (UID: \"70baf97d-1c3a-4106-9754-9369d381faf1\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" Sep 30 16:22:18 crc kubenswrapper[4796]: I0930 16:22:18.869596 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gdxx\" (UniqueName: \"kubernetes.io/projected/70baf97d-1c3a-4106-9754-9369d381faf1-kube-api-access-6gdxx\") pod \"nmstate-operator-5d6f6cfd66-jpgzs\" (UID: \"70baf97d-1c3a-4106-9754-9369d381faf1\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" Sep 30 16:22:18 crc kubenswrapper[4796]: I0930 16:22:18.953620 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" Sep 30 16:22:19 crc kubenswrapper[4796]: I0930 16:22:19.420350 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs"] Sep 30 16:22:20 crc kubenswrapper[4796]: I0930 16:22:20.080635 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" event={"ID":"70baf97d-1c3a-4106-9754-9369d381faf1","Type":"ContainerStarted","Data":"40237c0bf5dd021a18917a3a9f8cc6dab58d487ab35d2886d0ea300ab7f939e5"} Sep 30 16:22:22 crc kubenswrapper[4796]: I0930 16:22:22.099270 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" event={"ID":"70baf97d-1c3a-4106-9754-9369d381faf1","Type":"ContainerStarted","Data":"636edead58f12d9c5b86473426c5fac996393b6151deaab81216ed7698e90d72"} Sep 30 16:22:22 crc kubenswrapper[4796]: I0930 16:22:22.125022 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-jpgzs" podStartSLOduration=3.360609628 podStartE2EDuration="5.124958084s" podCreationTimestamp="2025-09-30 16:22:17 +0000 UTC" firstStartedPulling="2025-09-30 16:22:19.436010541 +0000 UTC m=+631.449289068" lastFinishedPulling="2025-09-30 16:22:21.200358967 +0000 UTC m=+633.213637524" observedRunningTime="2025-09-30 16:22:22.122189223 +0000 UTC m=+634.135467750" watchObservedRunningTime="2025-09-30 16:22:22.124958084 +0000 UTC m=+634.138236611" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.418675 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.419823 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.423926 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-p5b2d" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.460145 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.463041 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-hpgrl"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.463714 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.471443 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-9xkld"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.472455 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.478893 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.491312 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-9xkld"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.512598 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkxb\" (UniqueName: \"kubernetes.io/projected/541be51b-57e9-42d8-b0f7-02a1276c910f-kube-api-access-lbkxb\") pod \"nmstate-metrics-58fcddf996-9fgrw\" (UID: \"541be51b-57e9-42d8-b0f7-02a1276c910f\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.613779 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/90f4fdf8-d73e-47f1-9e33-41f58baf256b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.614203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqgs9\" (UniqueName: \"kubernetes.io/projected/90f4fdf8-d73e-47f1-9e33-41f58baf256b-kube-api-access-xqgs9\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.614255 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkxb\" (UniqueName: \"kubernetes.io/projected/541be51b-57e9-42d8-b0f7-02a1276c910f-kube-api-access-lbkxb\") pod \"nmstate-metrics-58fcddf996-9fgrw\" (UID: \"541be51b-57e9-42d8-b0f7-02a1276c910f\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.614289 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-ovs-socket\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.614316 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-dbus-socket\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.614338 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-nmstate-lock\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.614414 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw5ht\" (UniqueName: \"kubernetes.io/projected/bf0cf03b-b835-403a-83c5-dd587a330d94-kube-api-access-qw5ht\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.637997 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkxb\" (UniqueName: \"kubernetes.io/projected/541be51b-57e9-42d8-b0f7-02a1276c910f-kube-api-access-lbkxb\") pod \"nmstate-metrics-58fcddf996-9fgrw\" (UID: \"541be51b-57e9-42d8-b0f7-02a1276c910f\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.686968 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.687993 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.690787 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dxzhk" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.691225 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.691500 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.693202 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.716183 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-ovs-socket\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.716244 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-dbus-socket\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.716274 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-nmstate-lock\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.716340 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw5ht\" (UniqueName: \"kubernetes.io/projected/bf0cf03b-b835-403a-83c5-dd587a330d94-kube-api-access-qw5ht\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.716403 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/90f4fdf8-d73e-47f1-9e33-41f58baf256b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.716432 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqgs9\" (UniqueName: \"kubernetes.io/projected/90f4fdf8-d73e-47f1-9e33-41f58baf256b-kube-api-access-xqgs9\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.717073 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-nmstate-lock\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.717087 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-ovs-socket\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: E0930 16:22:28.717318 4796 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Sep 30 16:22:28 crc kubenswrapper[4796]: E0930 16:22:28.717439 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90f4fdf8-d73e-47f1-9e33-41f58baf256b-tls-key-pair podName:90f4fdf8-d73e-47f1-9e33-41f58baf256b nodeName:}" failed. No retries permitted until 2025-09-30 16:22:29.217414115 +0000 UTC m=+641.230692642 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/90f4fdf8-d73e-47f1-9e33-41f58baf256b-tls-key-pair") pod "nmstate-webhook-6d689559c5-9xkld" (UID: "90f4fdf8-d73e-47f1-9e33-41f58baf256b") : secret "openshift-nmstate-webhook" not found Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.717515 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bf0cf03b-b835-403a-83c5-dd587a330d94-dbus-socket\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.734381 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqgs9\" (UniqueName: \"kubernetes.io/projected/90f4fdf8-d73e-47f1-9e33-41f58baf256b-kube-api-access-xqgs9\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.739054 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.742881 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw5ht\" (UniqueName: \"kubernetes.io/projected/bf0cf03b-b835-403a-83c5-dd587a330d94-kube-api-access-qw5ht\") pod \"nmstate-handler-hpgrl\" (UID: \"bf0cf03b-b835-403a-83c5-dd587a330d94\") " pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.793370 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.818773 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/55affe5c-1791-49c1-9218-cbc57c6ee847-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.818842 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp7vm\" (UniqueName: \"kubernetes.io/projected/55affe5c-1791-49c1-9218-cbc57c6ee847-kube-api-access-fp7vm\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.818875 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55affe5c-1791-49c1-9218-cbc57c6ee847-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.872949 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-dcdccd4df-j6g7v"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.874218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.888479 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dcdccd4df-j6g7v"] Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.921018 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/55affe5c-1791-49c1-9218-cbc57c6ee847-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.921079 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp7vm\" (UniqueName: \"kubernetes.io/projected/55affe5c-1791-49c1-9218-cbc57c6ee847-kube-api-access-fp7vm\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.921116 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55affe5c-1791-49c1-9218-cbc57c6ee847-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: E0930 16:22:28.921278 4796 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Sep 30 16:22:28 crc kubenswrapper[4796]: E0930 16:22:28.921331 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55affe5c-1791-49c1-9218-cbc57c6ee847-plugin-serving-cert podName:55affe5c-1791-49c1-9218-cbc57c6ee847 nodeName:}" failed. No retries permitted until 2025-09-30 16:22:29.421316257 +0000 UTC m=+641.434594784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/55affe5c-1791-49c1-9218-cbc57c6ee847-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-96z6n" (UID: "55affe5c-1791-49c1-9218-cbc57c6ee847") : secret "plugin-serving-cert" not found Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.922372 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/55affe5c-1791-49c1-9218-cbc57c6ee847-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:28 crc kubenswrapper[4796]: I0930 16:22:28.949859 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp7vm\" (UniqueName: \"kubernetes.io/projected/55affe5c-1791-49c1-9218-cbc57c6ee847-kube-api-access-fp7vm\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022333 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-config\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022399 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4gb5\" (UniqueName: \"kubernetes.io/projected/0f5136a4-40b5-40bd-9460-ac12802d54b2-kube-api-access-c4gb5\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022479 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-trusted-ca-bundle\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022529 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-oauth-config\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022556 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-serving-cert\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022665 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-service-ca\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.022683 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-oauth-serving-cert\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.036346 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw"] Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123334 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-oauth-config\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123375 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-serving-cert\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123408 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-service-ca\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123428 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-oauth-serving-cert\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123468 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-config\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123494 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4gb5\" (UniqueName: \"kubernetes.io/projected/0f5136a4-40b5-40bd-9460-ac12802d54b2-kube-api-access-c4gb5\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.123531 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-trusted-ca-bundle\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.124837 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-trusted-ca-bundle\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.124922 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-oauth-serving-cert\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.125211 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-config\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.125243 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f5136a4-40b5-40bd-9460-ac12802d54b2-service-ca\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.127934 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-serving-cert\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.128019 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f5136a4-40b5-40bd-9460-ac12802d54b2-console-oauth-config\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.140768 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4gb5\" (UniqueName: \"kubernetes.io/projected/0f5136a4-40b5-40bd-9460-ac12802d54b2-kube-api-access-c4gb5\") pod \"console-dcdccd4df-j6g7v\" (UID: \"0f5136a4-40b5-40bd-9460-ac12802d54b2\") " pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.146441 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" event={"ID":"541be51b-57e9-42d8-b0f7-02a1276c910f","Type":"ContainerStarted","Data":"1d1a78f829d2dc32657957d6e22f957b113b536d47282e1865313f1fd8640593"} Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.151210 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hpgrl" event={"ID":"bf0cf03b-b835-403a-83c5-dd587a330d94","Type":"ContainerStarted","Data":"3c307299ab386e57247f38bc3c218def9f5bd6bb090e2dff2530323dab51f245"} Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.224486 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/90f4fdf8-d73e-47f1-9e33-41f58baf256b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.229955 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/90f4fdf8-d73e-47f1-9e33-41f58baf256b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-9xkld\" (UID: \"90f4fdf8-d73e-47f1-9e33-41f58baf256b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.234690 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.401500 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.430368 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55affe5c-1791-49c1-9218-cbc57c6ee847-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.437130 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55affe5c-1791-49c1-9218-cbc57c6ee847-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-96z6n\" (UID: \"55affe5c-1791-49c1-9218-cbc57c6ee847\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.529748 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dcdccd4df-j6g7v"] Sep 30 16:22:29 crc kubenswrapper[4796]: W0930 16:22:29.549588 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f5136a4_40b5_40bd_9460_ac12802d54b2.slice/crio-f1b88acabc0e0f9cec1e0bc8727fca90fedf19a396c2afad06cec90f7c25e21b WatchSource:0}: Error finding container f1b88acabc0e0f9cec1e0bc8727fca90fedf19a396c2afad06cec90f7c25e21b: Status 404 returned error can't find the container with id f1b88acabc0e0f9cec1e0bc8727fca90fedf19a396c2afad06cec90f7c25e21b Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.607181 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.669290 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-9xkld"] Sep 30 16:22:29 crc kubenswrapper[4796]: I0930 16:22:29.853804 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n"] Sep 30 16:22:30 crc kubenswrapper[4796]: I0930 16:22:30.159224 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dcdccd4df-j6g7v" event={"ID":"0f5136a4-40b5-40bd-9460-ac12802d54b2","Type":"ContainerStarted","Data":"f1b88acabc0e0f9cec1e0bc8727fca90fedf19a396c2afad06cec90f7c25e21b"} Sep 30 16:22:30 crc kubenswrapper[4796]: I0930 16:22:30.160784 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" event={"ID":"90f4fdf8-d73e-47f1-9e33-41f58baf256b","Type":"ContainerStarted","Data":"1b449e45a438f391fc04c6518bd64f2567f29e343f5459066977fbde1118e733"} Sep 30 16:22:30 crc kubenswrapper[4796]: I0930 16:22:30.162393 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" event={"ID":"55affe5c-1791-49c1-9218-cbc57c6ee847","Type":"ContainerStarted","Data":"f5847ee3719fd46b8351c4ee4692423f987d0917ebdcd060b9f259e43c35b008"} Sep 30 16:22:31 crc kubenswrapper[4796]: I0930 16:22:31.177304 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dcdccd4df-j6g7v" event={"ID":"0f5136a4-40b5-40bd-9460-ac12802d54b2","Type":"ContainerStarted","Data":"f7e989f9c13f525662c5e24e2f283ff80285417e064cb5cc8eb1b6368225f250"} Sep 30 16:22:31 crc kubenswrapper[4796]: I0930 16:22:31.200627 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-dcdccd4df-j6g7v" podStartSLOduration=3.200601051 podStartE2EDuration="3.200601051s" podCreationTimestamp="2025-09-30 16:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:22:31.198645144 +0000 UTC m=+643.211923681" watchObservedRunningTime="2025-09-30 16:22:31.200601051 +0000 UTC m=+643.213879578" Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.201749 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" event={"ID":"90f4fdf8-d73e-47f1-9e33-41f58baf256b","Type":"ContainerStarted","Data":"6605281d17c9d954ad4a7afaa8bcfe7008ee4dd2cefe020de7c2e3a44f780b11"} Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.202507 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.203657 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" event={"ID":"55affe5c-1791-49c1-9218-cbc57c6ee847","Type":"ContainerStarted","Data":"85f6b9a6871f690859a1fa7f9be4d96d368782d6f6d15d9f061e2cb0b058e0f3"} Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.204813 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" event={"ID":"541be51b-57e9-42d8-b0f7-02a1276c910f","Type":"ContainerStarted","Data":"335628b149b149f0350cb41b375f2ae2cb0aa939612a44e6ff2138ae20eb0c81"} Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.206049 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hpgrl" event={"ID":"bf0cf03b-b835-403a-83c5-dd587a330d94","Type":"ContainerStarted","Data":"9235d151c095f0415bdbb703f20c2528aeb15e2ddd53bd7c0aa6e2c9eef344f7"} Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.206164 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.218568 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" podStartSLOduration=2.186353833 podStartE2EDuration="5.218546807s" podCreationTimestamp="2025-09-30 16:22:28 +0000 UTC" firstStartedPulling="2025-09-30 16:22:29.687568119 +0000 UTC m=+641.700846646" lastFinishedPulling="2025-09-30 16:22:32.719761043 +0000 UTC m=+644.733039620" observedRunningTime="2025-09-30 16:22:33.217093715 +0000 UTC m=+645.230372262" watchObservedRunningTime="2025-09-30 16:22:33.218546807 +0000 UTC m=+645.231825354" Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.237003 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-96z6n" podStartSLOduration=2.380041886 podStartE2EDuration="5.236969268s" podCreationTimestamp="2025-09-30 16:22:28 +0000 UTC" firstStartedPulling="2025-09-30 16:22:29.863718217 +0000 UTC m=+641.876996754" lastFinishedPulling="2025-09-30 16:22:32.720645599 +0000 UTC m=+644.733924136" observedRunningTime="2025-09-30 16:22:33.233516196 +0000 UTC m=+645.246794723" watchObservedRunningTime="2025-09-30 16:22:33.236969268 +0000 UTC m=+645.250247795" Sep 30 16:22:33 crc kubenswrapper[4796]: I0930 16:22:33.255535 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-hpgrl" podStartSLOduration=1.331459571 podStartE2EDuration="5.255506822s" podCreationTimestamp="2025-09-30 16:22:28 +0000 UTC" firstStartedPulling="2025-09-30 16:22:28.833569163 +0000 UTC m=+640.846847680" lastFinishedPulling="2025-09-30 16:22:32.757616364 +0000 UTC m=+644.770894931" observedRunningTime="2025-09-30 16:22:33.253166073 +0000 UTC m=+645.266444620" watchObservedRunningTime="2025-09-30 16:22:33.255506822 +0000 UTC m=+645.268785379" Sep 30 16:22:35 crc kubenswrapper[4796]: I0930 16:22:35.225041 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" event={"ID":"541be51b-57e9-42d8-b0f7-02a1276c910f","Type":"ContainerStarted","Data":"cd2962b15aa246149751fed2aa2d98a5e272e2c3c39e1a967e16603d2af98fc6"} Sep 30 16:22:35 crc kubenswrapper[4796]: I0930 16:22:35.253465 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-9fgrw" podStartSLOduration=1.236693896 podStartE2EDuration="7.253432526s" podCreationTimestamp="2025-09-30 16:22:28 +0000 UTC" firstStartedPulling="2025-09-30 16:22:29.049102336 +0000 UTC m=+641.062380863" lastFinishedPulling="2025-09-30 16:22:35.065840966 +0000 UTC m=+647.079119493" observedRunningTime="2025-09-30 16:22:35.248212103 +0000 UTC m=+647.261490680" watchObservedRunningTime="2025-09-30 16:22:35.253432526 +0000 UTC m=+647.266711093" Sep 30 16:22:38 crc kubenswrapper[4796]: I0930 16:22:38.842517 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-hpgrl" Sep 30 16:22:39 crc kubenswrapper[4796]: I0930 16:22:39.235622 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:39 crc kubenswrapper[4796]: I0930 16:22:39.236180 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:39 crc kubenswrapper[4796]: I0930 16:22:39.243396 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:39 crc kubenswrapper[4796]: I0930 16:22:39.272016 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-dcdccd4df-j6g7v" Sep 30 16:22:39 crc kubenswrapper[4796]: I0930 16:22:39.346404 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cch5s"] Sep 30 16:22:49 crc kubenswrapper[4796]: I0930 16:22:49.411433 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-9xkld" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.396916 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-cch5s" podUID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" containerName="console" containerID="cri-o://990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf" gracePeriod=15 Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.803736 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cch5s_fc0a9b5c-9f56-41b0-8a45-aa5752b2da80/console/0.log" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.804100 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863167 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-service-ca\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863225 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-oauth-serving-cert\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863289 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-trusted-ca-bundle\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863391 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-oauth-config\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863459 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-config\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863478 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kzpb\" (UniqueName: \"kubernetes.io/projected/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-kube-api-access-7kzpb\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.863563 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-serving-cert\") pod \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\" (UID: \"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80\") " Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.864547 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-config" (OuterVolumeSpecName: "console-config") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.864584 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.864996 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-service-ca" (OuterVolumeSpecName: "service-ca") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.865392 4796 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.865417 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.865429 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.865582 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.871109 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-kube-api-access-7kzpb" (OuterVolumeSpecName: "kube-api-access-7kzpb") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "kube-api-access-7kzpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.871436 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.871886 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" (UID: "fc0a9b5c-9f56-41b0-8a45-aa5752b2da80"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.966857 4796 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.966951 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kzpb\" (UniqueName: \"kubernetes.io/projected/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-kube-api-access-7kzpb\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.966974 4796 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:04 crc kubenswrapper[4796]: I0930 16:23:04.967024 4796 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.467441 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cch5s_fc0a9b5c-9f56-41b0-8a45-aa5752b2da80/console/0.log" Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.469188 4796 generic.go:334] "Generic (PLEG): container finished" podID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" containerID="990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf" exitCode=2 Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.469238 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cch5s" Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.469273 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cch5s" event={"ID":"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80","Type":"ContainerDied","Data":"990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf"} Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.469390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cch5s" event={"ID":"fc0a9b5c-9f56-41b0-8a45-aa5752b2da80","Type":"ContainerDied","Data":"99c5a05428eae90dafbc3c3ab185d699edbcf71df1b1b5a1f08fa48b08c4584e"} Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.469444 4796 scope.go:117] "RemoveContainer" containerID="990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf" Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.496811 4796 scope.go:117] "RemoveContainer" containerID="990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf" Sep 30 16:23:05 crc kubenswrapper[4796]: E0930 16:23:05.497689 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf\": container with ID starting with 990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf not found: ID does not exist" containerID="990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf" Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.497742 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf"} err="failed to get container status \"990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf\": rpc error: code = NotFound desc = could not find container \"990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf\": container with ID starting with 990ab0c91d4d967d62ac904069facf5e00ae31f69f290f8e9bc4b7a923b879cf not found: ID does not exist" Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.503362 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cch5s"] Sep 30 16:23:05 crc kubenswrapper[4796]: I0930 16:23:05.506661 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-cch5s"] Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.567846 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd"] Sep 30 16:23:06 crc kubenswrapper[4796]: E0930 16:23:06.568283 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" containerName="console" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.568305 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" containerName="console" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.568505 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" containerName="console" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.569976 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.573414 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.584005 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd"] Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.694329 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jd4d\" (UniqueName: \"kubernetes.io/projected/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-kube-api-access-5jd4d\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.694619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.694750 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.743488 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0a9b5c-9f56-41b0-8a45-aa5752b2da80" path="/var/lib/kubelet/pods/fc0a9b5c-9f56-41b0-8a45-aa5752b2da80/volumes" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.796394 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.796453 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.796911 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.797064 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.797452 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jd4d\" (UniqueName: \"kubernetes.io/projected/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-kube-api-access-5jd4d\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.820122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jd4d\" (UniqueName: \"kubernetes.io/projected/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-kube-api-access-5jd4d\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:06 crc kubenswrapper[4796]: I0930 16:23:06.896005 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:07 crc kubenswrapper[4796]: I0930 16:23:07.317901 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd"] Sep 30 16:23:07 crc kubenswrapper[4796]: I0930 16:23:07.491904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" event={"ID":"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64","Type":"ContainerStarted","Data":"8d47e7c11a23756ea2f297e254ab6c59d02b22e5370420736aefb5f2592c7711"} Sep 30 16:23:07 crc kubenswrapper[4796]: I0930 16:23:07.491946 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" event={"ID":"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64","Type":"ContainerStarted","Data":"b5a631051f9f8bc3e93d0415f6f5c63e1d27e5551559ff5c8a3e04f83438ae2a"} Sep 30 16:23:08 crc kubenswrapper[4796]: I0930 16:23:08.502096 4796 generic.go:334] "Generic (PLEG): container finished" podID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerID="8d47e7c11a23756ea2f297e254ab6c59d02b22e5370420736aefb5f2592c7711" exitCode=0 Sep 30 16:23:08 crc kubenswrapper[4796]: I0930 16:23:08.502487 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" event={"ID":"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64","Type":"ContainerDied","Data":"8d47e7c11a23756ea2f297e254ab6c59d02b22e5370420736aefb5f2592c7711"} Sep 30 16:23:10 crc kubenswrapper[4796]: I0930 16:23:10.517425 4796 generic.go:334] "Generic (PLEG): container finished" podID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerID="195d32152a860bdb8ccc46de49fe5301d6f75ea2dcea8dbbd78246a9a5aefac8" exitCode=0 Sep 30 16:23:10 crc kubenswrapper[4796]: I0930 16:23:10.517517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" event={"ID":"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64","Type":"ContainerDied","Data":"195d32152a860bdb8ccc46de49fe5301d6f75ea2dcea8dbbd78246a9a5aefac8"} Sep 30 16:23:11 crc kubenswrapper[4796]: I0930 16:23:11.528367 4796 generic.go:334] "Generic (PLEG): container finished" podID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerID="ea5cd8b099a16e6eea4c4ad7bb084aef0fae67b511547ca9bcbf826fc1321e39" exitCode=0 Sep 30 16:23:11 crc kubenswrapper[4796]: I0930 16:23:11.528443 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" event={"ID":"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64","Type":"ContainerDied","Data":"ea5cd8b099a16e6eea4c4ad7bb084aef0fae67b511547ca9bcbf826fc1321e39"} Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.802473 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.882007 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-util\") pod \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.882084 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-bundle\") pod \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.882117 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jd4d\" (UniqueName: \"kubernetes.io/projected/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-kube-api-access-5jd4d\") pod \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\" (UID: \"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64\") " Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.885486 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-bundle" (OuterVolumeSpecName: "bundle") pod "1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" (UID: "1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.895303 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-kube-api-access-5jd4d" (OuterVolumeSpecName: "kube-api-access-5jd4d") pod "1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" (UID: "1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64"). InnerVolumeSpecName "kube-api-access-5jd4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.897431 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-util" (OuterVolumeSpecName: "util") pod "1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" (UID: "1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.984174 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.984243 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:12 crc kubenswrapper[4796]: I0930 16:23:12.984333 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jd4d\" (UniqueName: \"kubernetes.io/projected/1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64-kube-api-access-5jd4d\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:13 crc kubenswrapper[4796]: I0930 16:23:13.542155 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" event={"ID":"1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64","Type":"ContainerDied","Data":"b5a631051f9f8bc3e93d0415f6f5c63e1d27e5551559ff5c8a3e04f83438ae2a"} Sep 30 16:23:13 crc kubenswrapper[4796]: I0930 16:23:13.542203 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5a631051f9f8bc3e93d0415f6f5c63e1d27e5551559ff5c8a3e04f83438ae2a" Sep 30 16:23:13 crc kubenswrapper[4796]: I0930 16:23:13.542208 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.653268 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp"] Sep 30 16:23:21 crc kubenswrapper[4796]: E0930 16:23:21.654219 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="extract" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.654231 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="extract" Sep 30 16:23:21 crc kubenswrapper[4796]: E0930 16:23:21.654252 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="pull" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.654258 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="pull" Sep 30 16:23:21 crc kubenswrapper[4796]: E0930 16:23:21.654271 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="util" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.654277 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="util" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.654389 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64" containerName="extract" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.654781 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.657300 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.657307 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.657414 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.657671 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-h4mc5" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.658183 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.704763 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp"] Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.802784 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmfl5\" (UniqueName: \"kubernetes.io/projected/184ca539-6770-44bf-8461-bbde8fce05e1-kube-api-access-gmfl5\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.802886 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/184ca539-6770-44bf-8461-bbde8fce05e1-apiservice-cert\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.802943 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/184ca539-6770-44bf-8461-bbde8fce05e1-webhook-cert\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.883000 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w"] Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.883768 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.885955 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-78pnp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.886870 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.887123 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.901083 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w"] Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.904221 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmfl5\" (UniqueName: \"kubernetes.io/projected/184ca539-6770-44bf-8461-bbde8fce05e1-kube-api-access-gmfl5\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.904286 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/184ca539-6770-44bf-8461-bbde8fce05e1-apiservice-cert\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.904323 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/184ca539-6770-44bf-8461-bbde8fce05e1-webhook-cert\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.916382 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/184ca539-6770-44bf-8461-bbde8fce05e1-apiservice-cert\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.923034 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/184ca539-6770-44bf-8461-bbde8fce05e1-webhook-cert\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.933100 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmfl5\" (UniqueName: \"kubernetes.io/projected/184ca539-6770-44bf-8461-bbde8fce05e1-kube-api-access-gmfl5\") pod \"metallb-operator-controller-manager-588f64bf97-2tdzp\" (UID: \"184ca539-6770-44bf-8461-bbde8fce05e1\") " pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:21 crc kubenswrapper[4796]: I0930 16:23:21.972913 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.005751 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81e041f0-90d3-44a2-b3bd-d15212a55e24-webhook-cert\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.006343 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cwqd\" (UniqueName: \"kubernetes.io/projected/81e041f0-90d3-44a2-b3bd-d15212a55e24-kube-api-access-9cwqd\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.006375 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81e041f0-90d3-44a2-b3bd-d15212a55e24-apiservice-cert\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.109641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cwqd\" (UniqueName: \"kubernetes.io/projected/81e041f0-90d3-44a2-b3bd-d15212a55e24-kube-api-access-9cwqd\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.109689 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81e041f0-90d3-44a2-b3bd-d15212a55e24-apiservice-cert\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.109754 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81e041f0-90d3-44a2-b3bd-d15212a55e24-webhook-cert\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.119351 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81e041f0-90d3-44a2-b3bd-d15212a55e24-apiservice-cert\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.129352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81e041f0-90d3-44a2-b3bd-d15212a55e24-webhook-cert\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.133537 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cwqd\" (UniqueName: \"kubernetes.io/projected/81e041f0-90d3-44a2-b3bd-d15212a55e24-kube-api-access-9cwqd\") pod \"metallb-operator-webhook-server-54894f57bf-kfz9w\" (UID: \"81e041f0-90d3-44a2-b3bd-d15212a55e24\") " pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.200889 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.245848 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp"] Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.463973 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w"] Sep 30 16:23:22 crc kubenswrapper[4796]: W0930 16:23:22.468624 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81e041f0_90d3_44a2_b3bd_d15212a55e24.slice/crio-824096098b165c74e459fb7b6ec21a4a49c825dbb19ebda96589518387d730bd WatchSource:0}: Error finding container 824096098b165c74e459fb7b6ec21a4a49c825dbb19ebda96589518387d730bd: Status 404 returned error can't find the container with id 824096098b165c74e459fb7b6ec21a4a49c825dbb19ebda96589518387d730bd Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.597991 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" event={"ID":"81e041f0-90d3-44a2-b3bd-d15212a55e24","Type":"ContainerStarted","Data":"824096098b165c74e459fb7b6ec21a4a49c825dbb19ebda96589518387d730bd"} Sep 30 16:23:22 crc kubenswrapper[4796]: I0930 16:23:22.599037 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" event={"ID":"184ca539-6770-44bf-8461-bbde8fce05e1","Type":"ContainerStarted","Data":"b674561e222160dd57e3a9fb0b613568a1918df926be6c06c3e8d697febb7e2e"} Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.288534 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjzqr"] Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.289511 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" podUID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" containerName="controller-manager" containerID="cri-o://312c8723b3ee39245f8973360e13912b5368bd365df91ea6e13ed34bd0d453c4" gracePeriod=30 Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.306427 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn"] Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.306707 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" podUID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" containerName="route-controller-manager" containerID="cri-o://577a42dcb7c824000b52c2a69168d07b54bc585e5f27b9683c37a785afb67e98" gracePeriod=30 Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.661157 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" event={"ID":"184ca539-6770-44bf-8461-bbde8fce05e1","Type":"ContainerStarted","Data":"6f485b249c068321e6790c2a45e64692724fdb4372dd8a9a67ba0baef1ab785c"} Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.661755 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.664167 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" event={"ID":"81e041f0-90d3-44a2-b3bd-d15212a55e24","Type":"ContainerStarted","Data":"5aa82312bbd62494f207552f145c691fc70f538ef476dbdb0cdb3e7c4631471e"} Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.664958 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.666459 4796 generic.go:334] "Generic (PLEG): container finished" podID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" containerID="577a42dcb7c824000b52c2a69168d07b54bc585e5f27b9683c37a785afb67e98" exitCode=0 Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.666546 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" event={"ID":"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6","Type":"ContainerDied","Data":"577a42dcb7c824000b52c2a69168d07b54bc585e5f27b9683c37a785afb67e98"} Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.667706 4796 generic.go:334] "Generic (PLEG): container finished" podID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" containerID="312c8723b3ee39245f8973360e13912b5368bd365df91ea6e13ed34bd0d453c4" exitCode=0 Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.667739 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" event={"ID":"5271ef96-e4c6-49a4-b75b-b7ddf035402a","Type":"ContainerDied","Data":"312c8723b3ee39245f8973360e13912b5368bd365df91ea6e13ed34bd0d453c4"} Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.694241 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" podStartSLOduration=2.479496535 podStartE2EDuration="7.694212238s" podCreationTimestamp="2025-09-30 16:23:21 +0000 UTC" firstStartedPulling="2025-09-30 16:23:22.27322486 +0000 UTC m=+694.286503377" lastFinishedPulling="2025-09-30 16:23:27.487940543 +0000 UTC m=+699.501219080" observedRunningTime="2025-09-30 16:23:28.684233806 +0000 UTC m=+700.697512333" watchObservedRunningTime="2025-09-30 16:23:28.694212238 +0000 UTC m=+700.707490765" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.754604 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.758850 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.775574 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" podStartSLOduration=2.7333660760000003 podStartE2EDuration="7.775546338s" podCreationTimestamp="2025-09-30 16:23:21 +0000 UTC" firstStartedPulling="2025-09-30 16:23:22.471199995 +0000 UTC m=+694.484478522" lastFinishedPulling="2025-09-30 16:23:27.513380257 +0000 UTC m=+699.526658784" observedRunningTime="2025-09-30 16:23:28.719320623 +0000 UTC m=+700.732599150" watchObservedRunningTime="2025-09-30 16:23:28.775546338 +0000 UTC m=+700.788824875" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819039 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krvrl\" (UniqueName: \"kubernetes.io/projected/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-kube-api-access-krvrl\") pod \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819110 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-client-ca\") pod \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819192 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-config\") pod \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819215 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-serving-cert\") pod \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\" (UID: \"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819285 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-client-ca\") pod \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819316 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5271ef96-e4c6-49a4-b75b-b7ddf035402a-serving-cert\") pod \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819336 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-config\") pod \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819379 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-proxy-ca-bundles\") pod \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.819402 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd4j8\" (UniqueName: \"kubernetes.io/projected/5271ef96-e4c6-49a4-b75b-b7ddf035402a-kube-api-access-jd4j8\") pod \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\" (UID: \"5271ef96-e4c6-49a4-b75b-b7ddf035402a\") " Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.820893 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5271ef96-e4c6-49a4-b75b-b7ddf035402a" (UID: "5271ef96-e4c6-49a4-b75b-b7ddf035402a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.823437 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-client-ca" (OuterVolumeSpecName: "client-ca") pod "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" (UID: "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.823744 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-config" (OuterVolumeSpecName: "config") pod "5271ef96-e4c6-49a4-b75b-b7ddf035402a" (UID: "5271ef96-e4c6-49a4-b75b-b7ddf035402a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.823904 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-client-ca" (OuterVolumeSpecName: "client-ca") pod "5271ef96-e4c6-49a4-b75b-b7ddf035402a" (UID: "5271ef96-e4c6-49a4-b75b-b7ddf035402a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.824069 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-config" (OuterVolumeSpecName: "config") pod "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" (UID: "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.828723 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" (UID: "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.828752 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5271ef96-e4c6-49a4-b75b-b7ddf035402a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5271ef96-e4c6-49a4-b75b-b7ddf035402a" (UID: "5271ef96-e4c6-49a4-b75b-b7ddf035402a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.828943 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-kube-api-access-krvrl" (OuterVolumeSpecName: "kube-api-access-krvrl") pod "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" (UID: "89bad259-b74e-4bd2-b8cd-64b59ccc6ae6"). InnerVolumeSpecName "kube-api-access-krvrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.841970 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5271ef96-e4c6-49a4-b75b-b7ddf035402a-kube-api-access-jd4j8" (OuterVolumeSpecName: "kube-api-access-jd4j8") pod "5271ef96-e4c6-49a4-b75b-b7ddf035402a" (UID: "5271ef96-e4c6-49a4-b75b-b7ddf035402a"). InnerVolumeSpecName "kube-api-access-jd4j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920767 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920807 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920820 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920829 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920839 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5271ef96-e4c6-49a4-b75b-b7ddf035402a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920847 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920855 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5271ef96-e4c6-49a4-b75b-b7ddf035402a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920865 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd4j8\" (UniqueName: \"kubernetes.io/projected/5271ef96-e4c6-49a4-b75b-b7ddf035402a-kube-api-access-jd4j8\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:28 crc kubenswrapper[4796]: I0930 16:23:28.920875 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krvrl\" (UniqueName: \"kubernetes.io/projected/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6-kube-api-access-krvrl\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.676240 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.676572 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn" event={"ID":"89bad259-b74e-4bd2-b8cd-64b59ccc6ae6","Type":"ContainerDied","Data":"50d5e64161bac8ba3bdac4d599d27fc12a49edc371dbc589f930aa5851654033"} Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.676842 4796 scope.go:117] "RemoveContainer" containerID="577a42dcb7c824000b52c2a69168d07b54bc585e5f27b9683c37a785afb67e98" Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.681368 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" event={"ID":"5271ef96-e4c6-49a4-b75b-b7ddf035402a","Type":"ContainerDied","Data":"c11c76bb71f026108901068306d3e72c2b3484cc21a3982fb31d896f018944f4"} Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.681547 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjzqr" Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.701391 4796 scope.go:117] "RemoveContainer" containerID="312c8723b3ee39245f8973360e13912b5368bd365df91ea6e13ed34bd0d453c4" Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.732020 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn"] Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.738277 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6dvjn"] Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.750053 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjzqr"] Sep 30 16:23:29 crc kubenswrapper[4796]: I0930 16:23:29.751696 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjzqr"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.010776 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b769746cc-rx5v5"] Sep 30 16:23:30 crc kubenswrapper[4796]: E0930 16:23:30.011126 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" containerName="route-controller-manager" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.011147 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" containerName="route-controller-manager" Sep 30 16:23:30 crc kubenswrapper[4796]: E0930 16:23:30.011164 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" containerName="controller-manager" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.011172 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" containerName="controller-manager" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.011349 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" containerName="route-controller-manager" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.011380 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" containerName="controller-manager" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.011959 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.014938 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.015602 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.020692 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.020730 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.020777 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.020958 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577f787998-8j8br"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.022169 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.025337 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.026803 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.026899 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.027345 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.027536 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.027972 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.028775 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b769746cc-rx5v5"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.030832 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.036727 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.046291 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577f787998-8j8br"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.143590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r49x\" (UniqueName: \"kubernetes.io/projected/4f3f738d-5691-47cc-b892-39578936f872-kube-api-access-8r49x\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.143638 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-proxy-ca-bundles\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.143665 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgbj2\" (UniqueName: \"kubernetes.io/projected/276820ac-f980-48b9-b37b-9912e456cbfc-kube-api-access-cgbj2\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.143824 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f3f738d-5691-47cc-b892-39578936f872-serving-cert\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.143873 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-config\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.144001 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276820ac-f980-48b9-b37b-9912e456cbfc-serving-cert\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.144040 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-client-ca\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.144181 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-config\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.144226 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-client-ca\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245100 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r49x\" (UniqueName: \"kubernetes.io/projected/4f3f738d-5691-47cc-b892-39578936f872-kube-api-access-8r49x\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245183 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-proxy-ca-bundles\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245256 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgbj2\" (UniqueName: \"kubernetes.io/projected/276820ac-f980-48b9-b37b-9912e456cbfc-kube-api-access-cgbj2\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245299 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f3f738d-5691-47cc-b892-39578936f872-serving-cert\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245323 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-config\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245366 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276820ac-f980-48b9-b37b-9912e456cbfc-serving-cert\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245393 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-client-ca\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245452 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-config\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.245480 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-client-ca\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.247541 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-proxy-ca-bundles\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.248055 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-config\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.248578 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-client-ca\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.249735 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-client-ca\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.255800 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-config\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.257956 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f3f738d-5691-47cc-b892-39578936f872-serving-cert\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.267211 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276820ac-f980-48b9-b37b-9912e456cbfc-serving-cert\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.282674 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgbj2\" (UniqueName: \"kubernetes.io/projected/276820ac-f980-48b9-b37b-9912e456cbfc-kube-api-access-cgbj2\") pod \"route-controller-manager-577f787998-8j8br\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.286390 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r49x\" (UniqueName: \"kubernetes.io/projected/4f3f738d-5691-47cc-b892-39578936f872-kube-api-access-8r49x\") pod \"controller-manager-b769746cc-rx5v5\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.330339 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b769746cc-rx5v5"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.330804 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.333398 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577f787998-8j8br"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.333877 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.578560 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b769746cc-rx5v5"] Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.615279 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577f787998-8j8br"] Sep 30 16:23:30 crc kubenswrapper[4796]: W0930 16:23:30.624211 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod276820ac_f980_48b9_b37b_9912e456cbfc.slice/crio-32d3f6e2a49d5ef7d01a3933f59dbf3575255dd5eb835497e99f384d65194bfc WatchSource:0}: Error finding container 32d3f6e2a49d5ef7d01a3933f59dbf3575255dd5eb835497e99f384d65194bfc: Status 404 returned error can't find the container with id 32d3f6e2a49d5ef7d01a3933f59dbf3575255dd5eb835497e99f384d65194bfc Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.690217 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" event={"ID":"276820ac-f980-48b9-b37b-9912e456cbfc","Type":"ContainerStarted","Data":"32d3f6e2a49d5ef7d01a3933f59dbf3575255dd5eb835497e99f384d65194bfc"} Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.695567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" event={"ID":"4f3f738d-5691-47cc-b892-39578936f872","Type":"ContainerStarted","Data":"7c744591586af60a647c858664a8ea9c21d3727a10bc99ac5616c5d9512f0272"} Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.740544 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5271ef96-e4c6-49a4-b75b-b7ddf035402a" path="/var/lib/kubelet/pods/5271ef96-e4c6-49a4-b75b-b7ddf035402a/volumes" Sep 30 16:23:30 crc kubenswrapper[4796]: I0930 16:23:30.741354 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89bad259-b74e-4bd2-b8cd-64b59ccc6ae6" path="/var/lib/kubelet/pods/89bad259-b74e-4bd2-b8cd-64b59ccc6ae6/volumes" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.097386 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.097444 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.703764 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" event={"ID":"4f3f738d-5691-47cc-b892-39578936f872","Type":"ContainerStarted","Data":"da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f"} Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.704950 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.703819 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" podUID="4f3f738d-5691-47cc-b892-39578936f872" containerName="controller-manager" containerID="cri-o://da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f" gracePeriod=30 Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.705616 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" event={"ID":"276820ac-f980-48b9-b37b-9912e456cbfc","Type":"ContainerStarted","Data":"ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661"} Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.705921 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" podUID="276820ac-f980-48b9-b37b-9912e456cbfc" containerName="route-controller-manager" containerID="cri-o://ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661" gracePeriod=30 Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.706879 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.713401 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.714788 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.725817 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" podStartSLOduration=3.725795635 podStartE2EDuration="3.725795635s" podCreationTimestamp="2025-09-30 16:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:23:31.72562452 +0000 UTC m=+703.738903047" watchObservedRunningTime="2025-09-30 16:23:31.725795635 +0000 UTC m=+703.739074172" Sep 30 16:23:31 crc kubenswrapper[4796]: I0930 16:23:31.759656 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" podStartSLOduration=3.7596393949999998 podStartE2EDuration="3.759639395s" podCreationTimestamp="2025-09-30 16:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:23:31.755909226 +0000 UTC m=+703.769187753" watchObservedRunningTime="2025-09-30 16:23:31.759639395 +0000 UTC m=+703.772917922" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.132178 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.136565 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.161608 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8b778c78c-rj28c"] Sep 30 16:23:32 crc kubenswrapper[4796]: E0930 16:23:32.161863 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f3f738d-5691-47cc-b892-39578936f872" containerName="controller-manager" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.161878 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f3f738d-5691-47cc-b892-39578936f872" containerName="controller-manager" Sep 30 16:23:32 crc kubenswrapper[4796]: E0930 16:23:32.161904 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276820ac-f980-48b9-b37b-9912e456cbfc" containerName="route-controller-manager" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.161913 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="276820ac-f980-48b9-b37b-9912e456cbfc" containerName="route-controller-manager" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.162053 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="276820ac-f980-48b9-b37b-9912e456cbfc" containerName="route-controller-manager" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.162075 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f3f738d-5691-47cc-b892-39578936f872" containerName="controller-manager" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.162541 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170431 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r49x\" (UniqueName: \"kubernetes.io/projected/4f3f738d-5691-47cc-b892-39578936f872-kube-api-access-8r49x\") pod \"4f3f738d-5691-47cc-b892-39578936f872\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170482 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-client-ca\") pod \"276820ac-f980-48b9-b37b-9912e456cbfc\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170515 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f3f738d-5691-47cc-b892-39578936f872-serving-cert\") pod \"4f3f738d-5691-47cc-b892-39578936f872\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170558 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgbj2\" (UniqueName: \"kubernetes.io/projected/276820ac-f980-48b9-b37b-9912e456cbfc-kube-api-access-cgbj2\") pod \"276820ac-f980-48b9-b37b-9912e456cbfc\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170625 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276820ac-f980-48b9-b37b-9912e456cbfc-serving-cert\") pod \"276820ac-f980-48b9-b37b-9912e456cbfc\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170692 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-config\") pod \"276820ac-f980-48b9-b37b-9912e456cbfc\" (UID: \"276820ac-f980-48b9-b37b-9912e456cbfc\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170725 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-proxy-ca-bundles\") pod \"4f3f738d-5691-47cc-b892-39578936f872\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170778 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-client-ca\") pod \"4f3f738d-5691-47cc-b892-39578936f872\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.170801 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-config\") pod \"4f3f738d-5691-47cc-b892-39578936f872\" (UID: \"4f3f738d-5691-47cc-b892-39578936f872\") " Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.171948 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-client-ca" (OuterVolumeSpecName: "client-ca") pod "276820ac-f980-48b9-b37b-9912e456cbfc" (UID: "276820ac-f980-48b9-b37b-9912e456cbfc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.172400 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-config" (OuterVolumeSpecName: "config") pod "276820ac-f980-48b9-b37b-9912e456cbfc" (UID: "276820ac-f980-48b9-b37b-9912e456cbfc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.172483 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-config" (OuterVolumeSpecName: "config") pod "4f3f738d-5691-47cc-b892-39578936f872" (UID: "4f3f738d-5691-47cc-b892-39578936f872"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.172538 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-client-ca" (OuterVolumeSpecName: "client-ca") pod "4f3f738d-5691-47cc-b892-39578936f872" (UID: "4f3f738d-5691-47cc-b892-39578936f872"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.172563 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4f3f738d-5691-47cc-b892-39578936f872" (UID: "4f3f738d-5691-47cc-b892-39578936f872"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.181925 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3f738d-5691-47cc-b892-39578936f872-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4f3f738d-5691-47cc-b892-39578936f872" (UID: "4f3f738d-5691-47cc-b892-39578936f872"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.195217 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276820ac-f980-48b9-b37b-9912e456cbfc-kube-api-access-cgbj2" (OuterVolumeSpecName: "kube-api-access-cgbj2") pod "276820ac-f980-48b9-b37b-9912e456cbfc" (UID: "276820ac-f980-48b9-b37b-9912e456cbfc"). InnerVolumeSpecName "kube-api-access-cgbj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.198334 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f3f738d-5691-47cc-b892-39578936f872-kube-api-access-8r49x" (OuterVolumeSpecName: "kube-api-access-8r49x") pod "4f3f738d-5691-47cc-b892-39578936f872" (UID: "4f3f738d-5691-47cc-b892-39578936f872"). InnerVolumeSpecName "kube-api-access-8r49x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.202137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8b778c78c-rj28c"] Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.203197 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/276820ac-f980-48b9-b37b-9912e456cbfc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "276820ac-f980-48b9-b37b-9912e456cbfc" (UID: "276820ac-f980-48b9-b37b-9912e456cbfc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.272654 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-config\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.272714 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-proxy-ca-bundles\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.272785 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smwcx\" (UniqueName: \"kubernetes.io/projected/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-kube-api-access-smwcx\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.272829 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-serving-cert\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.272909 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-client-ca\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273021 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgbj2\" (UniqueName: \"kubernetes.io/projected/276820ac-f980-48b9-b37b-9912e456cbfc-kube-api-access-cgbj2\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273039 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276820ac-f980-48b9-b37b-9912e456cbfc-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273051 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273060 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273069 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273077 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f3f738d-5691-47cc-b892-39578936f872-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273085 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r49x\" (UniqueName: \"kubernetes.io/projected/4f3f738d-5691-47cc-b892-39578936f872-kube-api-access-8r49x\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273093 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276820ac-f980-48b9-b37b-9912e456cbfc-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.273101 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f3f738d-5691-47cc-b892-39578936f872-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.374624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smwcx\" (UniqueName: \"kubernetes.io/projected/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-kube-api-access-smwcx\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.374693 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-serving-cert\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.374750 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-client-ca\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.374815 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-config\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.374844 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-proxy-ca-bundles\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.376499 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-proxy-ca-bundles\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.379192 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-client-ca\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.380625 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-serving-cert\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.381076 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-config\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.398461 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smwcx\" (UniqueName: \"kubernetes.io/projected/b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc-kube-api-access-smwcx\") pod \"controller-manager-8b778c78c-rj28c\" (UID: \"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc\") " pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.482952 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.699315 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8b778c78c-rj28c"] Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.724529 4796 generic.go:334] "Generic (PLEG): container finished" podID="276820ac-f980-48b9-b37b-9912e456cbfc" containerID="ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661" exitCode=0 Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.724612 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" event={"ID":"276820ac-f980-48b9-b37b-9912e456cbfc","Type":"ContainerDied","Data":"ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661"} Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.724643 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" event={"ID":"276820ac-f980-48b9-b37b-9912e456cbfc","Type":"ContainerDied","Data":"32d3f6e2a49d5ef7d01a3933f59dbf3575255dd5eb835497e99f384d65194bfc"} Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.724663 4796 scope.go:117] "RemoveContainer" containerID="ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.725010 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577f787998-8j8br" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.736408 4796 generic.go:334] "Generic (PLEG): container finished" podID="4f3f738d-5691-47cc-b892-39578936f872" containerID="da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f" exitCode=0 Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.736759 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.756063 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" event={"ID":"4f3f738d-5691-47cc-b892-39578936f872","Type":"ContainerDied","Data":"da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f"} Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.756109 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b769746cc-rx5v5" event={"ID":"4f3f738d-5691-47cc-b892-39578936f872","Type":"ContainerDied","Data":"7c744591586af60a647c858664a8ea9c21d3727a10bc99ac5616c5d9512f0272"} Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.763281 4796 scope.go:117] "RemoveContainer" containerID="ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661" Sep 30 16:23:32 crc kubenswrapper[4796]: E0930 16:23:32.768700 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661\": container with ID starting with ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661 not found: ID does not exist" containerID="ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.768754 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661"} err="failed to get container status \"ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661\": rpc error: code = NotFound desc = could not find container \"ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661\": container with ID starting with ecffb7cd0fc2e011fa93d19fe1a24e6d4ef68fa3df4805e5bef33911137e1661 not found: ID does not exist" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.768790 4796 scope.go:117] "RemoveContainer" containerID="da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.796047 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577f787998-8j8br"] Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.805098 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577f787998-8j8br"] Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.806405 4796 scope.go:117] "RemoveContainer" containerID="da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f" Sep 30 16:23:32 crc kubenswrapper[4796]: E0930 16:23:32.807117 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f\": container with ID starting with da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f not found: ID does not exist" containerID="da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.807184 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f"} err="failed to get container status \"da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f\": rpc error: code = NotFound desc = could not find container \"da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f\": container with ID starting with da81adb000694cec851d1cbdd78e7bb91f77ebc11fcda3dff057807fe468825f not found: ID does not exist" Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.816266 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b769746cc-rx5v5"] Sep 30 16:23:32 crc kubenswrapper[4796]: I0930 16:23:32.816500 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b769746cc-rx5v5"] Sep 30 16:23:33 crc kubenswrapper[4796]: I0930 16:23:33.748710 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" event={"ID":"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc","Type":"ContainerStarted","Data":"e02306b9594667fdd81c26adeb44c04cbf4a6286e4ce9de55335ce2e9c31fc07"} Sep 30 16:23:33 crc kubenswrapper[4796]: I0930 16:23:33.749315 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:33 crc kubenswrapper[4796]: I0930 16:23:33.749333 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" event={"ID":"b0a4fa50-9cee-4e6f-9053-b06ae8e92fbc","Type":"ContainerStarted","Data":"4ae68d8355df048a77e4af4d8c04872193817bfa532c5f42b3521b4f3fa76c99"} Sep 30 16:23:33 crc kubenswrapper[4796]: I0930 16:23:33.756906 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" Sep 30 16:23:33 crc kubenswrapper[4796]: I0930 16:23:33.778015 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-8b778c78c-rj28c" podStartSLOduration=3.777995917 podStartE2EDuration="3.777995917s" podCreationTimestamp="2025-09-30 16:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:23:33.776777562 +0000 UTC m=+705.790056119" watchObservedRunningTime="2025-09-30 16:23:33.777995917 +0000 UTC m=+705.791274444" Sep 30 16:23:33 crc kubenswrapper[4796]: I0930 16:23:33.974098 4796 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 16:23:34 crc kubenswrapper[4796]: I0930 16:23:34.739684 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="276820ac-f980-48b9-b37b-9912e456cbfc" path="/var/lib/kubelet/pods/276820ac-f980-48b9-b37b-9912e456cbfc/volumes" Sep 30 16:23:34 crc kubenswrapper[4796]: I0930 16:23:34.740220 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f3f738d-5691-47cc-b892-39578936f872" path="/var/lib/kubelet/pods/4f3f738d-5691-47cc-b892-39578936f872/volumes" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.008525 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v"] Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.009227 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.011640 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.011748 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.011756 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.012195 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.012242 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.012300 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.025891 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v"] Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.108972 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-config\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.109041 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42mbv\" (UniqueName: \"kubernetes.io/projected/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-kube-api-access-42mbv\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.109100 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-client-ca\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.109147 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-serving-cert\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.211070 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-serving-cert\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.211158 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-config\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.211218 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42mbv\" (UniqueName: \"kubernetes.io/projected/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-kube-api-access-42mbv\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.211640 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-client-ca\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.212382 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-client-ca\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.212744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-config\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.217830 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-serving-cert\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.228572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42mbv\" (UniqueName: \"kubernetes.io/projected/30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad-kube-api-access-42mbv\") pod \"route-controller-manager-7d86d5fddd-jt87v\" (UID: \"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad\") " pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.336478 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:35 crc kubenswrapper[4796]: I0930 16:23:35.764361 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v"] Sep 30 16:23:35 crc kubenswrapper[4796]: W0930 16:23:35.772372 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30eff2ec_51c1_4868_99bb_6ac1c6c0d2ad.slice/crio-b86bc3d2ecd766a1f3bcf6ec8d2b3d3206ca7dea93433f6091b96e6f674dcb5b WatchSource:0}: Error finding container b86bc3d2ecd766a1f3bcf6ec8d2b3d3206ca7dea93433f6091b96e6f674dcb5b: Status 404 returned error can't find the container with id b86bc3d2ecd766a1f3bcf6ec8d2b3d3206ca7dea93433f6091b96e6f674dcb5b Sep 30 16:23:36 crc kubenswrapper[4796]: I0930 16:23:36.766276 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" event={"ID":"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad","Type":"ContainerStarted","Data":"94c0aa7af78c472b34206ab2f614d54f28242b0f80e884b4ed95c77cc49f5e42"} Sep 30 16:23:36 crc kubenswrapper[4796]: I0930 16:23:36.766630 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:36 crc kubenswrapper[4796]: I0930 16:23:36.766642 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" event={"ID":"30eff2ec-51c1-4868-99bb-6ac1c6c0d2ad","Type":"ContainerStarted","Data":"b86bc3d2ecd766a1f3bcf6ec8d2b3d3206ca7dea93433f6091b96e6f674dcb5b"} Sep 30 16:23:36 crc kubenswrapper[4796]: I0930 16:23:36.770827 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" Sep 30 16:23:36 crc kubenswrapper[4796]: I0930 16:23:36.789481 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7d86d5fddd-jt87v" podStartSLOduration=6.789459805 podStartE2EDuration="6.789459805s" podCreationTimestamp="2025-09-30 16:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:23:36.784443548 +0000 UTC m=+708.797722065" watchObservedRunningTime="2025-09-30 16:23:36.789459805 +0000 UTC m=+708.802738332" Sep 30 16:23:42 crc kubenswrapper[4796]: I0930 16:23:42.209737 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-54894f57bf-kfz9w" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.783190 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-788g5"] Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.785284 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.801165 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-788g5"] Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.897610 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-utilities\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.897665 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-catalog-content\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.897715 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmt5x\" (UniqueName: \"kubernetes.io/projected/c0b99a74-7284-477e-bc14-1878002df974-kube-api-access-dmt5x\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.999230 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmt5x\" (UniqueName: \"kubernetes.io/projected/c0b99a74-7284-477e-bc14-1878002df974-kube-api-access-dmt5x\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.999359 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-utilities\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:00 crc kubenswrapper[4796]: I0930 16:24:00.999387 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-catalog-content\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:00.999961 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-utilities\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.000012 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-catalog-content\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.021472 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmt5x\" (UniqueName: \"kubernetes.io/projected/c0b99a74-7284-477e-bc14-1878002df974-kube-api-access-dmt5x\") pod \"redhat-operators-788g5\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.097136 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.097214 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.116819 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.571496 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-788g5"] Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.938596 4796 generic.go:334] "Generic (PLEG): container finished" podID="c0b99a74-7284-477e-bc14-1878002df974" containerID="a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2" exitCode=0 Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.938665 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerDied","Data":"a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2"} Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.938701 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerStarted","Data":"ce914fd76a111e7fd3703a1c1086d12331d6af1ddefdf46cad4e998fb3d1fb96"} Sep 30 16:24:01 crc kubenswrapper[4796]: I0930 16:24:01.976150 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-588f64bf97-2tdzp" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.794057 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-v2psf"] Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.796773 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.798576 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.798822 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-np9gb" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.799371 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.804846 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr"] Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.805613 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.807606 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.818102 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr"] Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930288 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d03550b5-47a9-41ae-ad9b-1cc7d57c3335-cert\") pod \"frr-k8s-webhook-server-5478bdb765-bq9nr\" (UID: \"d03550b5-47a9-41ae-ad9b-1cc7d57c3335\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930361 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-conf\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930384 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-sockets\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930399 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-reloader\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930420 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930444 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics-certs\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930505 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg969\" (UniqueName: \"kubernetes.io/projected/d09a2502-5fca-4442-9f3b-f245b82ffd77-kube-api-access-rg969\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930528 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls9lj\" (UniqueName: \"kubernetes.io/projected/d03550b5-47a9-41ae-ad9b-1cc7d57c3335-kube-api-access-ls9lj\") pod \"frr-k8s-webhook-server-5478bdb765-bq9nr\" (UID: \"d03550b5-47a9-41ae-ad9b-1cc7d57c3335\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.930579 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-startup\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.936025 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tl8x6"] Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.940963 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tl8x6" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.945196 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.945419 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-m5b48" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.952744 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.952750 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.957400 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-z8rjn"] Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.958304 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.965140 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 30 16:24:02 crc kubenswrapper[4796]: I0930 16:24:02.970275 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-z8rjn"] Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.031932 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg969\" (UniqueName: \"kubernetes.io/projected/d09a2502-5fca-4442-9f3b-f245b82ffd77-kube-api-access-rg969\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032007 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls9lj\" (UniqueName: \"kubernetes.io/projected/d03550b5-47a9-41ae-ad9b-1cc7d57c3335-kube-api-access-ls9lj\") pod \"frr-k8s-webhook-server-5478bdb765-bq9nr\" (UID: \"d03550b5-47a9-41ae-ad9b-1cc7d57c3335\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032063 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-metrics-certs\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032103 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-785nz\" (UniqueName: \"kubernetes.io/projected/7b713269-c9b4-4619-9f57-7b932bcc2977-kube-api-access-785nz\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032133 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-startup\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032170 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-metrics-certs\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032193 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d03550b5-47a9-41ae-ad9b-1cc7d57c3335-cert\") pod \"frr-k8s-webhook-server-5478bdb765-bq9nr\" (UID: \"d03550b5-47a9-41ae-ad9b-1cc7d57c3335\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x42v\" (UniqueName: \"kubernetes.io/projected/5dd558be-ca13-43fc-9548-0ec92e3a2b69-kube-api-access-2x42v\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032458 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-cert\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032511 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5dd558be-ca13-43fc-9548-0ec92e3a2b69-metallb-excludel2\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032633 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-sockets\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032668 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-conf\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032720 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-reloader\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032790 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032819 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.032868 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics-certs\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.033112 4796 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.033130 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-sockets\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.033175 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-startup\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.033185 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics-certs podName:d09a2502-5fca-4442-9f3b-f245b82ffd77 nodeName:}" failed. No retries permitted until 2025-09-30 16:24:03.533139051 +0000 UTC m=+735.546417578 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics-certs") pod "frr-k8s-v2psf" (UID: "d09a2502-5fca-4442-9f3b-f245b82ffd77") : secret "frr-k8s-certs-secret" not found Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.033425 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-reloader\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.033518 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-frr-conf\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.033728 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.049168 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d03550b5-47a9-41ae-ad9b-1cc7d57c3335-cert\") pod \"frr-k8s-webhook-server-5478bdb765-bq9nr\" (UID: \"d03550b5-47a9-41ae-ad9b-1cc7d57c3335\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.051323 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg969\" (UniqueName: \"kubernetes.io/projected/d09a2502-5fca-4442-9f3b-f245b82ffd77-kube-api-access-rg969\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.051828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls9lj\" (UniqueName: \"kubernetes.io/projected/d03550b5-47a9-41ae-ad9b-1cc7d57c3335-kube-api-access-ls9lj\") pod \"frr-k8s-webhook-server-5478bdb765-bq9nr\" (UID: \"d03550b5-47a9-41ae-ad9b-1cc7d57c3335\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.133796 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.133910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-metrics-certs\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.133953 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-785nz\" (UniqueName: \"kubernetes.io/projected/7b713269-c9b4-4619-9f57-7b932bcc2977-kube-api-access-785nz\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.134011 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-metrics-certs\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.134041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x42v\" (UniqueName: \"kubernetes.io/projected/5dd558be-ca13-43fc-9548-0ec92e3a2b69-kube-api-access-2x42v\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.134064 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-cert\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.134087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5dd558be-ca13-43fc-9548-0ec92e3a2b69-metallb-excludel2\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.134117 4796 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.134181 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-metrics-certs podName:7b713269-c9b4-4619-9f57-7b932bcc2977 nodeName:}" failed. No retries permitted until 2025-09-30 16:24:03.634162829 +0000 UTC m=+735.647441356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-metrics-certs") pod "controller-5d688f5ffc-z8rjn" (UID: "7b713269-c9b4-4619-9f57-7b932bcc2977") : secret "controller-certs-secret" not found Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.134549 4796 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.134619 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist podName:5dd558be-ca13-43fc-9548-0ec92e3a2b69 nodeName:}" failed. No retries permitted until 2025-09-30 16:24:03.634603742 +0000 UTC m=+735.647882269 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist") pod "speaker-tl8x6" (UID: "5dd558be-ca13-43fc-9548-0ec92e3a2b69") : secret "metallb-memberlist" not found Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.134960 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5dd558be-ca13-43fc-9548-0ec92e3a2b69-metallb-excludel2\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.138382 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.138744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-metrics-certs\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.149429 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-cert\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.152566 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x42v\" (UniqueName: \"kubernetes.io/projected/5dd558be-ca13-43fc-9548-0ec92e3a2b69-kube-api-access-2x42v\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.162667 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-785nz\" (UniqueName: \"kubernetes.io/projected/7b713269-c9b4-4619-9f57-7b932bcc2977-kube-api-access-785nz\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.200311 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.539761 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics-certs\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.552154 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a2502-5fca-4442-9f3b-f245b82ffd77-metrics-certs\") pod \"frr-k8s-v2psf\" (UID: \"d09a2502-5fca-4442-9f3b-f245b82ffd77\") " pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.643710 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.643772 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-metrics-certs\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.643910 4796 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 16:24:03 crc kubenswrapper[4796]: E0930 16:24:03.644048 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist podName:5dd558be-ca13-43fc-9548-0ec92e3a2b69 nodeName:}" failed. No retries permitted until 2025-09-30 16:24:04.644017945 +0000 UTC m=+736.657296502 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist") pod "speaker-tl8x6" (UID: "5dd558be-ca13-43fc-9548-0ec92e3a2b69") : secret "metallb-memberlist" not found Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.652486 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b713269-c9b4-4619-9f57-7b932bcc2977-metrics-certs\") pod \"controller-5d688f5ffc-z8rjn\" (UID: \"7b713269-c9b4-4619-9f57-7b932bcc2977\") " pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.694947 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr"] Sep 30 16:24:03 crc kubenswrapper[4796]: W0930 16:24:03.706440 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd03550b5_47a9_41ae_ad9b_1cc7d57c3335.slice/crio-f9c09fcd1cad68619bcff8cfd11995b1c1d1827da137ec5428eb35437df52beb WatchSource:0}: Error finding container f9c09fcd1cad68619bcff8cfd11995b1c1d1827da137ec5428eb35437df52beb: Status 404 returned error can't find the container with id f9c09fcd1cad68619bcff8cfd11995b1c1d1827da137ec5428eb35437df52beb Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.720450 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.877798 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.951304 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerStarted","Data":"f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf"} Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.954464 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" event={"ID":"d03550b5-47a9-41ae-ad9b-1cc7d57c3335","Type":"ContainerStarted","Data":"f9c09fcd1cad68619bcff8cfd11995b1c1d1827da137ec5428eb35437df52beb"} Sep 30 16:24:03 crc kubenswrapper[4796]: I0930 16:24:03.957634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"0274365e2633a92be54d536a94b73f0004e0dc779deb07b2b46bab2d814f164f"} Sep 30 16:24:04 crc kubenswrapper[4796]: I0930 16:24:04.353350 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-z8rjn"] Sep 30 16:24:04 crc kubenswrapper[4796]: W0930 16:24:04.361897 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b713269_c9b4_4619_9f57_7b932bcc2977.slice/crio-5f94afa23ca536f90030abf261b5f0499b8b954cfbeb99d063a62e0f1e19706a WatchSource:0}: Error finding container 5f94afa23ca536f90030abf261b5f0499b8b954cfbeb99d063a62e0f1e19706a: Status 404 returned error can't find the container with id 5f94afa23ca536f90030abf261b5f0499b8b954cfbeb99d063a62e0f1e19706a Sep 30 16:24:04 crc kubenswrapper[4796]: I0930 16:24:04.661685 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:04 crc kubenswrapper[4796]: E0930 16:24:04.661870 4796 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 16:24:04 crc kubenswrapper[4796]: E0930 16:24:04.661938 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist podName:5dd558be-ca13-43fc-9548-0ec92e3a2b69 nodeName:}" failed. No retries permitted until 2025-09-30 16:24:06.661920475 +0000 UTC m=+738.675199002 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist") pod "speaker-tl8x6" (UID: "5dd558be-ca13-43fc-9548-0ec92e3a2b69") : secret "metallb-memberlist" not found Sep 30 16:24:04 crc kubenswrapper[4796]: I0930 16:24:04.965378 4796 generic.go:334] "Generic (PLEG): container finished" podID="c0b99a74-7284-477e-bc14-1878002df974" containerID="f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf" exitCode=0 Sep 30 16:24:04 crc kubenswrapper[4796]: I0930 16:24:04.965461 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerDied","Data":"f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf"} Sep 30 16:24:04 crc kubenswrapper[4796]: I0930 16:24:04.968334 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-z8rjn" event={"ID":"7b713269-c9b4-4619-9f57-7b932bcc2977","Type":"ContainerStarted","Data":"18132f1a152f75844a1bfba0fad14df77248d8117fd3000c5079ca36730107a0"} Sep 30 16:24:04 crc kubenswrapper[4796]: I0930 16:24:04.968377 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-z8rjn" event={"ID":"7b713269-c9b4-4619-9f57-7b932bcc2977","Type":"ContainerStarted","Data":"5f94afa23ca536f90030abf261b5f0499b8b954cfbeb99d063a62e0f1e19706a"} Sep 30 16:24:05 crc kubenswrapper[4796]: I0930 16:24:05.985784 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-z8rjn" event={"ID":"7b713269-c9b4-4619-9f57-7b932bcc2977","Type":"ContainerStarted","Data":"4d60a8f1f7e859258999c6e6aebedf7711424f16c95b8331e7845a350bf7bcef"} Sep 30 16:24:05 crc kubenswrapper[4796]: I0930 16:24:05.986236 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:06 crc kubenswrapper[4796]: I0930 16:24:06.010021 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-z8rjn" podStartSLOduration=4.009991697 podStartE2EDuration="4.009991697s" podCreationTimestamp="2025-09-30 16:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:24:06.004669399 +0000 UTC m=+738.017947926" watchObservedRunningTime="2025-09-30 16:24:06.009991697 +0000 UTC m=+738.023270224" Sep 30 16:24:06 crc kubenswrapper[4796]: I0930 16:24:06.694049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:06 crc kubenswrapper[4796]: I0930 16:24:06.701055 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5dd558be-ca13-43fc-9548-0ec92e3a2b69-memberlist\") pod \"speaker-tl8x6\" (UID: \"5dd558be-ca13-43fc-9548-0ec92e3a2b69\") " pod="metallb-system/speaker-tl8x6" Sep 30 16:24:06 crc kubenswrapper[4796]: I0930 16:24:06.857174 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tl8x6" Sep 30 16:24:06 crc kubenswrapper[4796]: W0930 16:24:06.890338 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dd558be_ca13_43fc_9548_0ec92e3a2b69.slice/crio-2e426ff3a9d0411ccceb946c90b8dfc5c3fc749a61d8f730ffd3fafb80e3746d WatchSource:0}: Error finding container 2e426ff3a9d0411ccceb946c90b8dfc5c3fc749a61d8f730ffd3fafb80e3746d: Status 404 returned error can't find the container with id 2e426ff3a9d0411ccceb946c90b8dfc5c3fc749a61d8f730ffd3fafb80e3746d Sep 30 16:24:06 crc kubenswrapper[4796]: I0930 16:24:06.999577 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tl8x6" event={"ID":"5dd558be-ca13-43fc-9548-0ec92e3a2b69","Type":"ContainerStarted","Data":"2e426ff3a9d0411ccceb946c90b8dfc5c3fc749a61d8f730ffd3fafb80e3746d"} Sep 30 16:24:08 crc kubenswrapper[4796]: I0930 16:24:08.008188 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tl8x6" event={"ID":"5dd558be-ca13-43fc-9548-0ec92e3a2b69","Type":"ContainerStarted","Data":"7117b5bf850e33bc160b8db1450f50cf2f436fa9193d8eeb4832dc3e58b1d198"} Sep 30 16:24:09 crc kubenswrapper[4796]: I0930 16:24:09.023276 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tl8x6" event={"ID":"5dd558be-ca13-43fc-9548-0ec92e3a2b69","Type":"ContainerStarted","Data":"aaaf08175c3561435e1fa3e98ae780eb1faaacfc5ad23a9303beff5563d6b883"} Sep 30 16:24:09 crc kubenswrapper[4796]: I0930 16:24:09.023741 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tl8x6" Sep 30 16:24:09 crc kubenswrapper[4796]: I0930 16:24:09.049408 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tl8x6" podStartSLOduration=7.049389071 podStartE2EDuration="7.049389071s" podCreationTimestamp="2025-09-30 16:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:24:09.048426123 +0000 UTC m=+741.061704730" watchObservedRunningTime="2025-09-30 16:24:09.049389071 +0000 UTC m=+741.062667598" Sep 30 16:24:13 crc kubenswrapper[4796]: I0930 16:24:13.061290 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerStarted","Data":"e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04"} Sep 30 16:24:14 crc kubenswrapper[4796]: I0930 16:24:14.916923 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h677s"] Sep 30 16:24:14 crc kubenswrapper[4796]: I0930 16:24:14.918021 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:14 crc kubenswrapper[4796]: I0930 16:24:14.927996 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h677s"] Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.026393 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-utilities\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.026492 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-catalog-content\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.026575 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gqb9\" (UniqueName: \"kubernetes.io/projected/acd12a25-37ad-441d-afb2-02d140314b77-kube-api-access-2gqb9\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.127939 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-utilities\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.128041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-catalog-content\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.128060 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gqb9\" (UniqueName: \"kubernetes.io/projected/acd12a25-37ad-441d-afb2-02d140314b77-kube-api-access-2gqb9\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.128560 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-utilities\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.128617 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-catalog-content\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.151247 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gqb9\" (UniqueName: \"kubernetes.io/projected/acd12a25-37ad-441d-afb2-02d140314b77-kube-api-access-2gqb9\") pod \"redhat-marketplace-h677s\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:15 crc kubenswrapper[4796]: I0930 16:24:15.267714 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:16 crc kubenswrapper[4796]: I0930 16:24:16.105449 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-788g5" podStartSLOduration=7.938323238 podStartE2EDuration="16.105420058s" podCreationTimestamp="2025-09-30 16:24:00 +0000 UTC" firstStartedPulling="2025-09-30 16:24:01.94042276 +0000 UTC m=+733.953701287" lastFinishedPulling="2025-09-30 16:24:10.10751953 +0000 UTC m=+742.120798107" observedRunningTime="2025-09-30 16:24:16.098502164 +0000 UTC m=+748.111780701" watchObservedRunningTime="2025-09-30 16:24:16.105420058 +0000 UTC m=+748.118698625" Sep 30 16:24:21 crc kubenswrapper[4796]: I0930 16:24:21.117076 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:21 crc kubenswrapper[4796]: I0930 16:24:21.118142 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:21 crc kubenswrapper[4796]: I0930 16:24:21.178776 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.091128 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.091642 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:cp-frr-files,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4,Command:[/bin/sh -c cp -rLf /tmp/frr/* /etc/frr/],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:frr-startup,ReadOnly:false,MountPath:/tmp/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:frr-conf,ReadOnly:false,MountPath:/etc/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rg969,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*100,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*101,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-v2psf_metallb-system(d09a2502-5fca-4442-9f3b-f245b82ffd77): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.092869 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-v2psf" podUID="d09a2502-5fca-4442-9f3b-f245b82ffd77" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.124886 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4\\\"\"" pod="metallb-system/frr-k8s-v2psf" podUID="d09a2502-5fca-4442-9f3b-f245b82ffd77" Sep 30 16:24:22 crc kubenswrapper[4796]: I0930 16:24:22.168238 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h677s"] Sep 30 16:24:22 crc kubenswrapper[4796]: I0930 16:24:22.191418 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.304941 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.305157 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:frr-k8s-webhook-server,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4,Command:[/frr-k8s],Args:[--log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=$(NAMESPACE) --metrics-bind-address=:7572],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:monitoring,HostPort:0,ContainerPort:7572,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ls9lj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000730000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-webhook-server-5478bdb765-bq9nr_metallb-system(d03550b5-47a9-41ae-ad9b-1cc7d57c3335): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:24:22 crc kubenswrapper[4796]: E0930 16:24:22.306329 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" podUID="d03550b5-47a9-41ae-ad9b-1cc7d57c3335" Sep 30 16:24:22 crc kubenswrapper[4796]: I0930 16:24:22.415744 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-788g5"] Sep 30 16:24:23 crc kubenswrapper[4796]: I0930 16:24:23.135899 4796 generic.go:334] "Generic (PLEG): container finished" podID="acd12a25-37ad-441d-afb2-02d140314b77" containerID="9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525" exitCode=0 Sep 30 16:24:23 crc kubenswrapper[4796]: I0930 16:24:23.137104 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerDied","Data":"9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525"} Sep 30 16:24:23 crc kubenswrapper[4796]: I0930 16:24:23.137158 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerStarted","Data":"b9c329c06d929dbb39fc914e11e6c07bb1b7539509d873d4e51e55372bd0d961"} Sep 30 16:24:23 crc kubenswrapper[4796]: E0930 16:24:23.140172 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4\\\"\"" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" podUID="d03550b5-47a9-41ae-ad9b-1cc7d57c3335" Sep 30 16:24:23 crc kubenswrapper[4796]: I0930 16:24:23.883103 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-z8rjn" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.146345 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerStarted","Data":"2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d"} Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.146415 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-788g5" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="registry-server" containerID="cri-o://e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04" gracePeriod=2 Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.636723 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.681955 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-catalog-content\") pod \"c0b99a74-7284-477e-bc14-1878002df974\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.682126 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmt5x\" (UniqueName: \"kubernetes.io/projected/c0b99a74-7284-477e-bc14-1878002df974-kube-api-access-dmt5x\") pod \"c0b99a74-7284-477e-bc14-1878002df974\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.682160 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-utilities\") pod \"c0b99a74-7284-477e-bc14-1878002df974\" (UID: \"c0b99a74-7284-477e-bc14-1878002df974\") " Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.683140 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-utilities" (OuterVolumeSpecName: "utilities") pod "c0b99a74-7284-477e-bc14-1878002df974" (UID: "c0b99a74-7284-477e-bc14-1878002df974"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.689513 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b99a74-7284-477e-bc14-1878002df974-kube-api-access-dmt5x" (OuterVolumeSpecName: "kube-api-access-dmt5x") pod "c0b99a74-7284-477e-bc14-1878002df974" (UID: "c0b99a74-7284-477e-bc14-1878002df974"). InnerVolumeSpecName "kube-api-access-dmt5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.754719 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0b99a74-7284-477e-bc14-1878002df974" (UID: "c0b99a74-7284-477e-bc14-1878002df974"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.782951 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.783010 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmt5x\" (UniqueName: \"kubernetes.io/projected/c0b99a74-7284-477e-bc14-1878002df974-kube-api-access-dmt5x\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:24 crc kubenswrapper[4796]: I0930 16:24:24.783028 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b99a74-7284-477e-bc14-1878002df974-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.155859 4796 generic.go:334] "Generic (PLEG): container finished" podID="acd12a25-37ad-441d-afb2-02d140314b77" containerID="2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d" exitCode=0 Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.155962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerDied","Data":"2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d"} Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.158447 4796 generic.go:334] "Generic (PLEG): container finished" podID="c0b99a74-7284-477e-bc14-1878002df974" containerID="e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04" exitCode=0 Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.158496 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerDied","Data":"e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04"} Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.158554 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-788g5" event={"ID":"c0b99a74-7284-477e-bc14-1878002df974","Type":"ContainerDied","Data":"ce914fd76a111e7fd3703a1c1086d12331d6af1ddefdf46cad4e998fb3d1fb96"} Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.158563 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-788g5" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.158583 4796 scope.go:117] "RemoveContainer" containerID="e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.190694 4796 scope.go:117] "RemoveContainer" containerID="f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.214117 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-788g5"] Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.220927 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-788g5"] Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.230077 4796 scope.go:117] "RemoveContainer" containerID="a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.253610 4796 scope.go:117] "RemoveContainer" containerID="e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04" Sep 30 16:24:25 crc kubenswrapper[4796]: E0930 16:24:25.254194 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04\": container with ID starting with e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04 not found: ID does not exist" containerID="e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.254254 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04"} err="failed to get container status \"e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04\": rpc error: code = NotFound desc = could not find container \"e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04\": container with ID starting with e33306d913b80f22fd88cf004420617e67ca59e96a2881e62dafc2e58a40ed04 not found: ID does not exist" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.254290 4796 scope.go:117] "RemoveContainer" containerID="f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf" Sep 30 16:24:25 crc kubenswrapper[4796]: E0930 16:24:25.254971 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf\": container with ID starting with f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf not found: ID does not exist" containerID="f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.255042 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf"} err="failed to get container status \"f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf\": rpc error: code = NotFound desc = could not find container \"f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf\": container with ID starting with f6ec67d68311bae43d8af9b1a2d6e3a832e95de5eb690c9a04d67da9aea683cf not found: ID does not exist" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.255070 4796 scope.go:117] "RemoveContainer" containerID="a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2" Sep 30 16:24:25 crc kubenswrapper[4796]: E0930 16:24:25.255439 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2\": container with ID starting with a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2 not found: ID does not exist" containerID="a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2" Sep 30 16:24:25 crc kubenswrapper[4796]: I0930 16:24:25.255483 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2"} err="failed to get container status \"a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2\": rpc error: code = NotFound desc = could not find container \"a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2\": container with ID starting with a2be57f59502cac8f8071ba97dbd8bd7b6a621c5ea1acc81f9a09843ddff88b2 not found: ID does not exist" Sep 30 16:24:26 crc kubenswrapper[4796]: I0930 16:24:26.167630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerStarted","Data":"6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae"} Sep 30 16:24:26 crc kubenswrapper[4796]: I0930 16:24:26.191220 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h677s" podStartSLOduration=9.575918821 podStartE2EDuration="12.191198486s" podCreationTimestamp="2025-09-30 16:24:14 +0000 UTC" firstStartedPulling="2025-09-30 16:24:23.140073283 +0000 UTC m=+755.153351830" lastFinishedPulling="2025-09-30 16:24:25.755352958 +0000 UTC m=+757.768631495" observedRunningTime="2025-09-30 16:24:26.19067283 +0000 UTC m=+758.203951387" watchObservedRunningTime="2025-09-30 16:24:26.191198486 +0000 UTC m=+758.204477013" Sep 30 16:24:26 crc kubenswrapper[4796]: I0930 16:24:26.740936 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b99a74-7284-477e-bc14-1878002df974" path="/var/lib/kubelet/pods/c0b99a74-7284-477e-bc14-1878002df974/volumes" Sep 30 16:24:26 crc kubenswrapper[4796]: I0930 16:24:26.861641 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tl8x6" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.005315 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mczns"] Sep 30 16:24:30 crc kubenswrapper[4796]: E0930 16:24:30.006404 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="extract-utilities" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.006420 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="extract-utilities" Sep 30 16:24:30 crc kubenswrapper[4796]: E0930 16:24:30.006437 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="extract-content" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.006444 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="extract-content" Sep 30 16:24:30 crc kubenswrapper[4796]: E0930 16:24:30.006459 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="registry-server" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.006467 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="registry-server" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.006611 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b99a74-7284-477e-bc14-1878002df974" containerName="registry-server" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.007146 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.017544 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.018201 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.018202 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8x2nf" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.026211 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mczns"] Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.055608 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm8pq\" (UniqueName: \"kubernetes.io/projected/c7bd9f0c-706b-469e-8855-a5b9c29789a1-kube-api-access-lm8pq\") pod \"openstack-operator-index-mczns\" (UID: \"c7bd9f0c-706b-469e-8855-a5b9c29789a1\") " pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.156351 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm8pq\" (UniqueName: \"kubernetes.io/projected/c7bd9f0c-706b-469e-8855-a5b9c29789a1-kube-api-access-lm8pq\") pod \"openstack-operator-index-mczns\" (UID: \"c7bd9f0c-706b-469e-8855-a5b9c29789a1\") " pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.177382 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm8pq\" (UniqueName: \"kubernetes.io/projected/c7bd9f0c-706b-469e-8855-a5b9c29789a1-kube-api-access-lm8pq\") pod \"openstack-operator-index-mczns\" (UID: \"c7bd9f0c-706b-469e-8855-a5b9c29789a1\") " pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.349793 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:30 crc kubenswrapper[4796]: I0930 16:24:30.861051 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mczns"] Sep 30 16:24:31 crc kubenswrapper[4796]: I0930 16:24:31.097653 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:24:31 crc kubenswrapper[4796]: I0930 16:24:31.097758 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:24:31 crc kubenswrapper[4796]: I0930 16:24:31.097839 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:24:31 crc kubenswrapper[4796]: I0930 16:24:31.098684 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e292f0f61906ba72a5d11ef43cd6aabb03ee9b745e2fef2028fac6c22e38f4f"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:24:31 crc kubenswrapper[4796]: I0930 16:24:31.098765 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://8e292f0f61906ba72a5d11ef43cd6aabb03ee9b745e2fef2028fac6c22e38f4f" gracePeriod=600 Sep 30 16:24:31 crc kubenswrapper[4796]: I0930 16:24:31.209594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mczns" event={"ID":"c7bd9f0c-706b-469e-8855-a5b9c29789a1","Type":"ContainerStarted","Data":"cc7ff6516fca3729cf5b34f4ea5f59f86a8ff72e414729237b3eeb47ecc76b4c"} Sep 30 16:24:32 crc kubenswrapper[4796]: I0930 16:24:32.217516 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="8e292f0f61906ba72a5d11ef43cd6aabb03ee9b745e2fef2028fac6c22e38f4f" exitCode=0 Sep 30 16:24:32 crc kubenswrapper[4796]: I0930 16:24:32.217564 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"8e292f0f61906ba72a5d11ef43cd6aabb03ee9b745e2fef2028fac6c22e38f4f"} Sep 30 16:24:32 crc kubenswrapper[4796]: I0930 16:24:32.217594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"f80108aae73ea16e4bbeae96cdafd6d40fe314e16173a204b191d54c90a06ba4"} Sep 30 16:24:32 crc kubenswrapper[4796]: I0930 16:24:32.217613 4796 scope.go:117] "RemoveContainer" containerID="4ba0c7bae7193c789288a51223146b959225052eddbb4fa93010e8f66b95083d" Sep 30 16:24:34 crc kubenswrapper[4796]: I0930 16:24:34.215289 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mczns"] Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.026046 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-f5fb8"] Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.027391 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.028122 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lksrr\" (UniqueName: \"kubernetes.io/projected/3d4983f8-cba0-43ba-b756-0cab71ed31f8-kube-api-access-lksrr\") pod \"openstack-operator-index-f5fb8\" (UID: \"3d4983f8-cba0-43ba-b756-0cab71ed31f8\") " pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.032280 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f5fb8"] Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.130088 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lksrr\" (UniqueName: \"kubernetes.io/projected/3d4983f8-cba0-43ba-b756-0cab71ed31f8-kube-api-access-lksrr\") pod \"openstack-operator-index-f5fb8\" (UID: \"3d4983f8-cba0-43ba-b756-0cab71ed31f8\") " pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.152017 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lksrr\" (UniqueName: \"kubernetes.io/projected/3d4983f8-cba0-43ba-b756-0cab71ed31f8-kube-api-access-lksrr\") pod \"openstack-operator-index-f5fb8\" (UID: \"3d4983f8-cba0-43ba-b756-0cab71ed31f8\") " pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.239105 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mczns" event={"ID":"c7bd9f0c-706b-469e-8855-a5b9c29789a1","Type":"ContainerStarted","Data":"fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552"} Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.239345 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-mczns" podUID="c7bd9f0c-706b-469e-8855-a5b9c29789a1" containerName="registry-server" containerID="cri-o://fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552" gracePeriod=2 Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.262742 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mczns" podStartSLOduration=3.049176946 podStartE2EDuration="6.26272162s" podCreationTimestamp="2025-09-30 16:24:29 +0000 UTC" firstStartedPulling="2025-09-30 16:24:30.869902104 +0000 UTC m=+762.883180631" lastFinishedPulling="2025-09-30 16:24:34.083446778 +0000 UTC m=+766.096725305" observedRunningTime="2025-09-30 16:24:35.259904206 +0000 UTC m=+767.273182733" watchObservedRunningTime="2025-09-30 16:24:35.26272162 +0000 UTC m=+767.276000157" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.268828 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.268915 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.317502 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.352626 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.667557 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.793496 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f5fb8"] Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.843879 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm8pq\" (UniqueName: \"kubernetes.io/projected/c7bd9f0c-706b-469e-8855-a5b9c29789a1-kube-api-access-lm8pq\") pod \"c7bd9f0c-706b-469e-8855-a5b9c29789a1\" (UID: \"c7bd9f0c-706b-469e-8855-a5b9c29789a1\") " Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.849815 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7bd9f0c-706b-469e-8855-a5b9c29789a1-kube-api-access-lm8pq" (OuterVolumeSpecName: "kube-api-access-lm8pq") pod "c7bd9f0c-706b-469e-8855-a5b9c29789a1" (UID: "c7bd9f0c-706b-469e-8855-a5b9c29789a1"). InnerVolumeSpecName "kube-api-access-lm8pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:24:35 crc kubenswrapper[4796]: I0930 16:24:35.946178 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm8pq\" (UniqueName: \"kubernetes.io/projected/c7bd9f0c-706b-469e-8855-a5b9c29789a1-kube-api-access-lm8pq\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.245492 4796 generic.go:334] "Generic (PLEG): container finished" podID="c7bd9f0c-706b-469e-8855-a5b9c29789a1" containerID="fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552" exitCode=0 Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.245525 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mczns" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.245525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mczns" event={"ID":"c7bd9f0c-706b-469e-8855-a5b9c29789a1","Type":"ContainerDied","Data":"fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552"} Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.246186 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mczns" event={"ID":"c7bd9f0c-706b-469e-8855-a5b9c29789a1","Type":"ContainerDied","Data":"cc7ff6516fca3729cf5b34f4ea5f59f86a8ff72e414729237b3eeb47ecc76b4c"} Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.246206 4796 scope.go:117] "RemoveContainer" containerID="fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.250007 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f5fb8" event={"ID":"3d4983f8-cba0-43ba-b756-0cab71ed31f8","Type":"ContainerStarted","Data":"e2dda7effd1ce19291477fdeb99434d47cae05ad279de84284f6df17d1e0be7e"} Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.250056 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f5fb8" event={"ID":"3d4983f8-cba0-43ba-b756-0cab71ed31f8","Type":"ContainerStarted","Data":"3d783ba16e4ae36647aee863e75c327b39c21c66262fea4f466fa9b4fb0cb5b8"} Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.269214 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-f5fb8" podStartSLOduration=1.227363964 podStartE2EDuration="1.269194931s" podCreationTimestamp="2025-09-30 16:24:35 +0000 UTC" firstStartedPulling="2025-09-30 16:24:35.800501672 +0000 UTC m=+767.813780209" lastFinishedPulling="2025-09-30 16:24:35.842332649 +0000 UTC m=+767.855611176" observedRunningTime="2025-09-30 16:24:36.267730688 +0000 UTC m=+768.281009215" watchObservedRunningTime="2025-09-30 16:24:36.269194931 +0000 UTC m=+768.282473468" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.271423 4796 scope.go:117] "RemoveContainer" containerID="fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552" Sep 30 16:24:36 crc kubenswrapper[4796]: E0930 16:24:36.272578 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552\": container with ID starting with fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552 not found: ID does not exist" containerID="fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.272610 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552"} err="failed to get container status \"fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552\": rpc error: code = NotFound desc = could not find container \"fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552\": container with ID starting with fdff16768293bf4d801d40e02f49ff7ed112e37b7404b0a4f9866fa0c09e0552 not found: ID does not exist" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.282418 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mczns"] Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.286679 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-mczns"] Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.324211 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:36 crc kubenswrapper[4796]: I0930 16:24:36.747108 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7bd9f0c-706b-469e-8855-a5b9c29789a1" path="/var/lib/kubelet/pods/c7bd9f0c-706b-469e-8855-a5b9c29789a1/volumes" Sep 30 16:24:37 crc kubenswrapper[4796]: I0930 16:24:37.259186 4796 generic.go:334] "Generic (PLEG): container finished" podID="d09a2502-5fca-4442-9f3b-f245b82ffd77" containerID="82fd5bdd9d818050361552392d0ed9fdbf1873e84f6ce7d8bfb48a234420f4b0" exitCode=0 Sep 30 16:24:37 crc kubenswrapper[4796]: I0930 16:24:37.259257 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerDied","Data":"82fd5bdd9d818050361552392d0ed9fdbf1873e84f6ce7d8bfb48a234420f4b0"} Sep 30 16:24:37 crc kubenswrapper[4796]: I0930 16:24:37.266551 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" event={"ID":"d03550b5-47a9-41ae-ad9b-1cc7d57c3335","Type":"ContainerStarted","Data":"24d544414af19cfeb34d2c5c4192b1dbc3973601e1bd7856e2fee9d5cad78f60"} Sep 30 16:24:37 crc kubenswrapper[4796]: I0930 16:24:37.266858 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:37 crc kubenswrapper[4796]: I0930 16:24:37.308149 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" podStartSLOduration=2.169797812 podStartE2EDuration="35.308129983s" podCreationTimestamp="2025-09-30 16:24:02 +0000 UTC" firstStartedPulling="2025-09-30 16:24:03.710410528 +0000 UTC m=+735.723689095" lastFinishedPulling="2025-09-30 16:24:36.848742689 +0000 UTC m=+768.862021266" observedRunningTime="2025-09-30 16:24:37.305644329 +0000 UTC m=+769.318922856" watchObservedRunningTime="2025-09-30 16:24:37.308129983 +0000 UTC m=+769.321408500" Sep 30 16:24:38 crc kubenswrapper[4796]: I0930 16:24:38.272949 4796 generic.go:334] "Generic (PLEG): container finished" podID="d09a2502-5fca-4442-9f3b-f245b82ffd77" containerID="c19f00ac20809eab5bbbc495a4f50c74fd599048e02d8a696b2aeaa9b6c0f667" exitCode=0 Sep 30 16:24:38 crc kubenswrapper[4796]: I0930 16:24:38.273024 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerDied","Data":"c19f00ac20809eab5bbbc495a4f50c74fd599048e02d8a696b2aeaa9b6c0f667"} Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.215362 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h677s"] Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.215703 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h677s" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="registry-server" containerID="cri-o://6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae" gracePeriod=2 Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.282736 4796 generic.go:334] "Generic (PLEG): container finished" podID="d09a2502-5fca-4442-9f3b-f245b82ffd77" containerID="3ed3618517eabd3264331647427318558b98b6ba4b47633746bcf6aed841490c" exitCode=0 Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.282838 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerDied","Data":"3ed3618517eabd3264331647427318558b98b6ba4b47633746bcf6aed841490c"} Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.661221 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.707940 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-utilities\") pod \"acd12a25-37ad-441d-afb2-02d140314b77\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.708056 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-catalog-content\") pod \"acd12a25-37ad-441d-afb2-02d140314b77\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.708183 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gqb9\" (UniqueName: \"kubernetes.io/projected/acd12a25-37ad-441d-afb2-02d140314b77-kube-api-access-2gqb9\") pod \"acd12a25-37ad-441d-afb2-02d140314b77\" (UID: \"acd12a25-37ad-441d-afb2-02d140314b77\") " Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.708893 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-utilities" (OuterVolumeSpecName: "utilities") pod "acd12a25-37ad-441d-afb2-02d140314b77" (UID: "acd12a25-37ad-441d-afb2-02d140314b77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.718250 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd12a25-37ad-441d-afb2-02d140314b77-kube-api-access-2gqb9" (OuterVolumeSpecName: "kube-api-access-2gqb9") pod "acd12a25-37ad-441d-afb2-02d140314b77" (UID: "acd12a25-37ad-441d-afb2-02d140314b77"). InnerVolumeSpecName "kube-api-access-2gqb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.740349 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "acd12a25-37ad-441d-afb2-02d140314b77" (UID: "acd12a25-37ad-441d-afb2-02d140314b77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.810623 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gqb9\" (UniqueName: \"kubernetes.io/projected/acd12a25-37ad-441d-afb2-02d140314b77-kube-api-access-2gqb9\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.810649 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:39 crc kubenswrapper[4796]: I0930 16:24:39.811360 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd12a25-37ad-441d-afb2-02d140314b77-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.296503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"b6969d9b08ea5288bc777a1a87dc18dae734081d2f1ac0da2396880e523f626f"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.297015 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"22bb8c68caf20715895af1b2aa45edfed65861506dc26b7c8b7da5e65973dd25"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.297034 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"8b70a3f26fb535824463d1a8feb82c61281aee2ed5ae62d60cb2337240b6fb5e"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.297048 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"b385fecf44daf1f5b3384985b1b33a3647dbef7f6994fa0b3aad952ffacc4e61"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.297058 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"5e604f63e299e79c0a900e8ef203cb712437c0f0b98366586983232976f1b2c2"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.304700 4796 generic.go:334] "Generic (PLEG): container finished" podID="acd12a25-37ad-441d-afb2-02d140314b77" containerID="6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae" exitCode=0 Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.304749 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerDied","Data":"6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.304783 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h677s" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.304801 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h677s" event={"ID":"acd12a25-37ad-441d-afb2-02d140314b77","Type":"ContainerDied","Data":"b9c329c06d929dbb39fc914e11e6c07bb1b7539509d873d4e51e55372bd0d961"} Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.304825 4796 scope.go:117] "RemoveContainer" containerID="6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.335491 4796 scope.go:117] "RemoveContainer" containerID="2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.345972 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h677s"] Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.350909 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h677s"] Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.381345 4796 scope.go:117] "RemoveContainer" containerID="9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.398319 4796 scope.go:117] "RemoveContainer" containerID="6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae" Sep 30 16:24:40 crc kubenswrapper[4796]: E0930 16:24:40.398677 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae\": container with ID starting with 6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae not found: ID does not exist" containerID="6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.398712 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae"} err="failed to get container status \"6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae\": rpc error: code = NotFound desc = could not find container \"6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae\": container with ID starting with 6e8a0c00716ea00d9f1d978704a75d9473cba5a1b334d992fee45484c67f17ae not found: ID does not exist" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.398734 4796 scope.go:117] "RemoveContainer" containerID="2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d" Sep 30 16:24:40 crc kubenswrapper[4796]: E0930 16:24:40.398956 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d\": container with ID starting with 2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d not found: ID does not exist" containerID="2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.398990 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d"} err="failed to get container status \"2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d\": rpc error: code = NotFound desc = could not find container \"2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d\": container with ID starting with 2e456dfdf399cf55db2760369637a83acba0950c45b51aecd96c58a16d51322d not found: ID does not exist" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.399007 4796 scope.go:117] "RemoveContainer" containerID="9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525" Sep 30 16:24:40 crc kubenswrapper[4796]: E0930 16:24:40.399202 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525\": container with ID starting with 9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525 not found: ID does not exist" containerID="9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.399221 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525"} err="failed to get container status \"9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525\": rpc error: code = NotFound desc = could not find container \"9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525\": container with ID starting with 9902bf7cbfab4d00b90ff8e5a25f530564a40a8b7597fe4a5ea869ec398ea525 not found: ID does not exist" Sep 30 16:24:40 crc kubenswrapper[4796]: I0930 16:24:40.741491 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd12a25-37ad-441d-afb2-02d140314b77" path="/var/lib/kubelet/pods/acd12a25-37ad-441d-afb2-02d140314b77/volumes" Sep 30 16:24:41 crc kubenswrapper[4796]: I0930 16:24:41.325596 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2psf" event={"ID":"d09a2502-5fca-4442-9f3b-f245b82ffd77","Type":"ContainerStarted","Data":"75499907d74638b969876c338de46282a1de8b7c3d545a48e908c44e61279fc7"} Sep 30 16:24:41 crc kubenswrapper[4796]: I0930 16:24:41.328210 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:41 crc kubenswrapper[4796]: I0930 16:24:41.367868 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-v2psf" podStartSLOduration=6.408022156 podStartE2EDuration="39.367850688s" podCreationTimestamp="2025-09-30 16:24:02 +0000 UTC" firstStartedPulling="2025-09-30 16:24:03.888559436 +0000 UTC m=+735.901837973" lastFinishedPulling="2025-09-30 16:24:36.848387968 +0000 UTC m=+768.861666505" observedRunningTime="2025-09-30 16:24:41.367337253 +0000 UTC m=+773.380615790" watchObservedRunningTime="2025-09-30 16:24:41.367850688 +0000 UTC m=+773.381129215" Sep 30 16:24:43 crc kubenswrapper[4796]: I0930 16:24:43.721599 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:43 crc kubenswrapper[4796]: I0930 16:24:43.762653 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:45 crc kubenswrapper[4796]: I0930 16:24:45.354088 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:45 crc kubenswrapper[4796]: I0930 16:24:45.354618 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:45 crc kubenswrapper[4796]: I0930 16:24:45.406431 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:45 crc kubenswrapper[4796]: I0930 16:24:45.447994 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-f5fb8" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.086109 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9mwn7"] Sep 30 16:24:49 crc kubenswrapper[4796]: E0930 16:24:49.089162 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="extract-utilities" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.089311 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="extract-utilities" Sep 30 16:24:49 crc kubenswrapper[4796]: E0930 16:24:49.089427 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7bd9f0c-706b-469e-8855-a5b9c29789a1" containerName="registry-server" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.089549 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7bd9f0c-706b-469e-8855-a5b9c29789a1" containerName="registry-server" Sep 30 16:24:49 crc kubenswrapper[4796]: E0930 16:24:49.089652 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="extract-content" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.089744 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="extract-content" Sep 30 16:24:49 crc kubenswrapper[4796]: E0930 16:24:49.089890 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="registry-server" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.090105 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="registry-server" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.090950 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7bd9f0c-706b-469e-8855-a5b9c29789a1" containerName="registry-server" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.093082 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd12a25-37ad-441d-afb2-02d140314b77" containerName="registry-server" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.094769 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.098731 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9mwn7"] Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.262772 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-utilities\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.262828 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-catalog-content\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.262850 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxrnm\" (UniqueName: \"kubernetes.io/projected/e7a77ad7-bd47-46b0-b464-b96769d499f1-kube-api-access-nxrnm\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.364686 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-utilities\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.364746 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-catalog-content\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.364764 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxrnm\" (UniqueName: \"kubernetes.io/projected/e7a77ad7-bd47-46b0-b464-b96769d499f1-kube-api-access-nxrnm\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.365491 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-utilities\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.365700 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-catalog-content\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.389721 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxrnm\" (UniqueName: \"kubernetes.io/projected/e7a77ad7-bd47-46b0-b464-b96769d499f1-kube-api-access-nxrnm\") pod \"certified-operators-9mwn7\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.413569 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:49 crc kubenswrapper[4796]: I0930 16:24:49.920020 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9mwn7"] Sep 30 16:24:49 crc kubenswrapper[4796]: W0930 16:24:49.926676 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7a77ad7_bd47_46b0_b464_b96769d499f1.slice/crio-b7ab0f56eb5e47204b2d1b00e1d5bf1984830702704367cc958dd0dcb0e5497e WatchSource:0}: Error finding container b7ab0f56eb5e47204b2d1b00e1d5bf1984830702704367cc958dd0dcb0e5497e: Status 404 returned error can't find the container with id b7ab0f56eb5e47204b2d1b00e1d5bf1984830702704367cc958dd0dcb0e5497e Sep 30 16:24:50 crc kubenswrapper[4796]: I0930 16:24:50.404661 4796 generic.go:334] "Generic (PLEG): container finished" podID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerID="441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e" exitCode=0 Sep 30 16:24:50 crc kubenswrapper[4796]: I0930 16:24:50.404750 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mwn7" event={"ID":"e7a77ad7-bd47-46b0-b464-b96769d499f1","Type":"ContainerDied","Data":"441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e"} Sep 30 16:24:50 crc kubenswrapper[4796]: I0930 16:24:50.405172 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mwn7" event={"ID":"e7a77ad7-bd47-46b0-b464-b96769d499f1","Type":"ContainerStarted","Data":"b7ab0f56eb5e47204b2d1b00e1d5bf1984830702704367cc958dd0dcb0e5497e"} Sep 30 16:24:52 crc kubenswrapper[4796]: I0930 16:24:52.423402 4796 generic.go:334] "Generic (PLEG): container finished" podID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerID="5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa" exitCode=0 Sep 30 16:24:52 crc kubenswrapper[4796]: I0930 16:24:52.423707 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mwn7" event={"ID":"e7a77ad7-bd47-46b0-b464-b96769d499f1","Type":"ContainerDied","Data":"5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa"} Sep 30 16:24:53 crc kubenswrapper[4796]: I0930 16:24:53.209209 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-bq9nr" Sep 30 16:24:53 crc kubenswrapper[4796]: I0930 16:24:53.433795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mwn7" event={"ID":"e7a77ad7-bd47-46b0-b464-b96769d499f1","Type":"ContainerStarted","Data":"40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52"} Sep 30 16:24:53 crc kubenswrapper[4796]: I0930 16:24:53.461236 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9mwn7" podStartSLOduration=2.05383963 podStartE2EDuration="4.461214799s" podCreationTimestamp="2025-09-30 16:24:49 +0000 UTC" firstStartedPulling="2025-09-30 16:24:50.40769947 +0000 UTC m=+782.420977997" lastFinishedPulling="2025-09-30 16:24:52.815074639 +0000 UTC m=+784.828353166" observedRunningTime="2025-09-30 16:24:53.457449156 +0000 UTC m=+785.470727703" watchObservedRunningTime="2025-09-30 16:24:53.461214799 +0000 UTC m=+785.474493326" Sep 30 16:24:53 crc kubenswrapper[4796]: I0930 16:24:53.723914 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-v2psf" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.268946 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xln5d"] Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.270390 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.297013 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xln5d"] Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.451580 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d2e366d-8e88-4f1b-b2da-8db4ee157808-catalog-content\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.452088 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpvbw\" (UniqueName: \"kubernetes.io/projected/1d2e366d-8e88-4f1b-b2da-8db4ee157808-kube-api-access-tpvbw\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.452149 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d2e366d-8e88-4f1b-b2da-8db4ee157808-utilities\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.553075 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d2e366d-8e88-4f1b-b2da-8db4ee157808-utilities\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.553188 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d2e366d-8e88-4f1b-b2da-8db4ee157808-catalog-content\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.553233 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpvbw\" (UniqueName: \"kubernetes.io/projected/1d2e366d-8e88-4f1b-b2da-8db4ee157808-kube-api-access-tpvbw\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.553669 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d2e366d-8e88-4f1b-b2da-8db4ee157808-utilities\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.553773 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d2e366d-8e88-4f1b-b2da-8db4ee157808-catalog-content\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.579878 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpvbw\" (UniqueName: \"kubernetes.io/projected/1d2e366d-8e88-4f1b-b2da-8db4ee157808-kube-api-access-tpvbw\") pod \"community-operators-xln5d\" (UID: \"1d2e366d-8e88-4f1b-b2da-8db4ee157808\") " pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:55 crc kubenswrapper[4796]: I0930 16:24:55.594773 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:24:56 crc kubenswrapper[4796]: I0930 16:24:56.076093 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xln5d"] Sep 30 16:24:56 crc kubenswrapper[4796]: I0930 16:24:56.458523 4796 generic.go:334] "Generic (PLEG): container finished" podID="1d2e366d-8e88-4f1b-b2da-8db4ee157808" containerID="351b3f5042a85a760d228a083f6cc47fbd94124b908cac6a266cac23695e98e1" exitCode=0 Sep 30 16:24:56 crc kubenswrapper[4796]: I0930 16:24:56.458587 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xln5d" event={"ID":"1d2e366d-8e88-4f1b-b2da-8db4ee157808","Type":"ContainerDied","Data":"351b3f5042a85a760d228a083f6cc47fbd94124b908cac6a266cac23695e98e1"} Sep 30 16:24:56 crc kubenswrapper[4796]: I0930 16:24:56.458634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xln5d" event={"ID":"1d2e366d-8e88-4f1b-b2da-8db4ee157808","Type":"ContainerStarted","Data":"0bf522249abd2f0a1bece45a0a748ee078b0cd02215b4f50fb05062d28fc6fab"} Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.415111 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.415652 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.534470 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.579813 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.704114 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj"] Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.705593 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.712186 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-r8cb2" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.716796 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj"] Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.824730 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-bundle\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.825223 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-util\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.825341 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxz2x\" (UniqueName: \"kubernetes.io/projected/d5833467-cda7-456a-9f81-cd6b6b32a205-kube-api-access-pxz2x\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.926431 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxz2x\" (UniqueName: \"kubernetes.io/projected/d5833467-cda7-456a-9f81-cd6b6b32a205-kube-api-access-pxz2x\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.926572 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-bundle\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.926601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-util\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.929381 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-bundle\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.930179 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-util\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:24:59 crc kubenswrapper[4796]: I0930 16:24:59.949828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxz2x\" (UniqueName: \"kubernetes.io/projected/d5833467-cda7-456a-9f81-cd6b6b32a205-kube-api-access-pxz2x\") pod \"fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:25:00 crc kubenswrapper[4796]: I0930 16:25:00.026880 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:25:01 crc kubenswrapper[4796]: I0930 16:25:01.031342 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj"] Sep 30 16:25:01 crc kubenswrapper[4796]: W0930 16:25:01.047257 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5833467_cda7_456a_9f81_cd6b6b32a205.slice/crio-c2bb98a7513e6c772262c3cd47a58ebbefd6e760c8e7cff56cecdbb0efdf097c WatchSource:0}: Error finding container c2bb98a7513e6c772262c3cd47a58ebbefd6e760c8e7cff56cecdbb0efdf097c: Status 404 returned error can't find the container with id c2bb98a7513e6c772262c3cd47a58ebbefd6e760c8e7cff56cecdbb0efdf097c Sep 30 16:25:01 crc kubenswrapper[4796]: I0930 16:25:01.496102 4796 generic.go:334] "Generic (PLEG): container finished" podID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerID="4f3bfd7d22bf4ef4ec33744a215fbadb457481d94d89b2a08d2b2f269c6b3c3c" exitCode=0 Sep 30 16:25:01 crc kubenswrapper[4796]: I0930 16:25:01.496180 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" event={"ID":"d5833467-cda7-456a-9f81-cd6b6b32a205","Type":"ContainerDied","Data":"4f3bfd7d22bf4ef4ec33744a215fbadb457481d94d89b2a08d2b2f269c6b3c3c"} Sep 30 16:25:01 crc kubenswrapper[4796]: I0930 16:25:01.496213 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" event={"ID":"d5833467-cda7-456a-9f81-cd6b6b32a205","Type":"ContainerStarted","Data":"c2bb98a7513e6c772262c3cd47a58ebbefd6e760c8e7cff56cecdbb0efdf097c"} Sep 30 16:25:01 crc kubenswrapper[4796]: I0930 16:25:01.500187 4796 generic.go:334] "Generic (PLEG): container finished" podID="1d2e366d-8e88-4f1b-b2da-8db4ee157808" containerID="2144fd6b8961bf02fa887ad70b3d9d7985505aff66f7bf2810ca263b03fa86d3" exitCode=0 Sep 30 16:25:01 crc kubenswrapper[4796]: I0930 16:25:01.500218 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xln5d" event={"ID":"1d2e366d-8e88-4f1b-b2da-8db4ee157808","Type":"ContainerDied","Data":"2144fd6b8961bf02fa887ad70b3d9d7985505aff66f7bf2810ca263b03fa86d3"} Sep 30 16:25:02 crc kubenswrapper[4796]: I0930 16:25:02.512830 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xln5d" event={"ID":"1d2e366d-8e88-4f1b-b2da-8db4ee157808","Type":"ContainerStarted","Data":"ce024acda5d313877c7076de5ede1ff6dc2e183ced3903156a79b75c2d8a5fb3"} Sep 30 16:25:02 crc kubenswrapper[4796]: I0930 16:25:02.515433 4796 generic.go:334] "Generic (PLEG): container finished" podID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerID="7cd13af044ff5070f72fe9030dfcb8e04fa7a2236df30109524970a074eee816" exitCode=0 Sep 30 16:25:02 crc kubenswrapper[4796]: I0930 16:25:02.515508 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" event={"ID":"d5833467-cda7-456a-9f81-cd6b6b32a205","Type":"ContainerDied","Data":"7cd13af044ff5070f72fe9030dfcb8e04fa7a2236df30109524970a074eee816"} Sep 30 16:25:02 crc kubenswrapper[4796]: I0930 16:25:02.541656 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xln5d" podStartSLOduration=1.989638386 podStartE2EDuration="7.541632571s" podCreationTimestamp="2025-09-30 16:24:55 +0000 UTC" firstStartedPulling="2025-09-30 16:24:56.46082608 +0000 UTC m=+788.474104637" lastFinishedPulling="2025-09-30 16:25:02.012820285 +0000 UTC m=+794.026098822" observedRunningTime="2025-09-30 16:25:02.537841807 +0000 UTC m=+794.551120404" watchObservedRunningTime="2025-09-30 16:25:02.541632571 +0000 UTC m=+794.554911128" Sep 30 16:25:03 crc kubenswrapper[4796]: I0930 16:25:03.522711 4796 generic.go:334] "Generic (PLEG): container finished" podID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerID="0690da384262cba6d1b735dc19f9f640f44da21a65d1f64347ee3d09bb8ade3c" exitCode=0 Sep 30 16:25:03 crc kubenswrapper[4796]: I0930 16:25:03.522784 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" event={"ID":"d5833467-cda7-456a-9f81-cd6b6b32a205","Type":"ContainerDied","Data":"0690da384262cba6d1b735dc19f9f640f44da21a65d1f64347ee3d09bb8ade3c"} Sep 30 16:25:03 crc kubenswrapper[4796]: I0930 16:25:03.656206 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9mwn7"] Sep 30 16:25:03 crc kubenswrapper[4796]: I0930 16:25:03.656584 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9mwn7" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="registry-server" containerID="cri-o://40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52" gracePeriod=2 Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.106027 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.209957 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxrnm\" (UniqueName: \"kubernetes.io/projected/e7a77ad7-bd47-46b0-b464-b96769d499f1-kube-api-access-nxrnm\") pod \"e7a77ad7-bd47-46b0-b464-b96769d499f1\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.210042 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-utilities\") pod \"e7a77ad7-bd47-46b0-b464-b96769d499f1\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.210092 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-catalog-content\") pod \"e7a77ad7-bd47-46b0-b464-b96769d499f1\" (UID: \"e7a77ad7-bd47-46b0-b464-b96769d499f1\") " Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.215243 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-utilities" (OuterVolumeSpecName: "utilities") pod "e7a77ad7-bd47-46b0-b464-b96769d499f1" (UID: "e7a77ad7-bd47-46b0-b464-b96769d499f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.220441 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a77ad7-bd47-46b0-b464-b96769d499f1-kube-api-access-nxrnm" (OuterVolumeSpecName: "kube-api-access-nxrnm") pod "e7a77ad7-bd47-46b0-b464-b96769d499f1" (UID: "e7a77ad7-bd47-46b0-b464-b96769d499f1"). InnerVolumeSpecName "kube-api-access-nxrnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.257261 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7a77ad7-bd47-46b0-b464-b96769d499f1" (UID: "e7a77ad7-bd47-46b0-b464-b96769d499f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.311780 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxrnm\" (UniqueName: \"kubernetes.io/projected/e7a77ad7-bd47-46b0-b464-b96769d499f1-kube-api-access-nxrnm\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.311831 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.311844 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a77ad7-bd47-46b0-b464-b96769d499f1-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.530924 4796 generic.go:334] "Generic (PLEG): container finished" podID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerID="40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52" exitCode=0 Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.530995 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mwn7" event={"ID":"e7a77ad7-bd47-46b0-b464-b96769d499f1","Type":"ContainerDied","Data":"40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52"} Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.531476 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mwn7" event={"ID":"e7a77ad7-bd47-46b0-b464-b96769d499f1","Type":"ContainerDied","Data":"b7ab0f56eb5e47204b2d1b00e1d5bf1984830702704367cc958dd0dcb0e5497e"} Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.531504 4796 scope.go:117] "RemoveContainer" containerID="40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.531035 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mwn7" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.556528 4796 scope.go:117] "RemoveContainer" containerID="5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.565708 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9mwn7"] Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.568760 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9mwn7"] Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.597917 4796 scope.go:117] "RemoveContainer" containerID="441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.620292 4796 scope.go:117] "RemoveContainer" containerID="40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52" Sep 30 16:25:04 crc kubenswrapper[4796]: E0930 16:25:04.631716 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52\": container with ID starting with 40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52 not found: ID does not exist" containerID="40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.631841 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52"} err="failed to get container status \"40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52\": rpc error: code = NotFound desc = could not find container \"40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52\": container with ID starting with 40df38543bb57abf73a7d57c08d3279a434be81f9c8e5676899eab2e17566a52 not found: ID does not exist" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.631875 4796 scope.go:117] "RemoveContainer" containerID="5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa" Sep 30 16:25:04 crc kubenswrapper[4796]: E0930 16:25:04.632730 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa\": container with ID starting with 5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa not found: ID does not exist" containerID="5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.632763 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa"} err="failed to get container status \"5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa\": rpc error: code = NotFound desc = could not find container \"5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa\": container with ID starting with 5747d01c32233a62a9e57f5ea2499a3eec18418c4fc37d6f0fdf8eaa1ca92caa not found: ID does not exist" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.632786 4796 scope.go:117] "RemoveContainer" containerID="441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e" Sep 30 16:25:04 crc kubenswrapper[4796]: E0930 16:25:04.633177 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e\": container with ID starting with 441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e not found: ID does not exist" containerID="441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.633220 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e"} err="failed to get container status \"441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e\": rpc error: code = NotFound desc = could not find container \"441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e\": container with ID starting with 441fbfc5c230d797d247a6537bcb8932433c8e14c74022f49808d25e90f25b5e not found: ID does not exist" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.743294 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.745607 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" path="/var/lib/kubelet/pods/e7a77ad7-bd47-46b0-b464-b96769d499f1/volumes" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.819939 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxz2x\" (UniqueName: \"kubernetes.io/projected/d5833467-cda7-456a-9f81-cd6b6b32a205-kube-api-access-pxz2x\") pod \"d5833467-cda7-456a-9f81-cd6b6b32a205\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.820029 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-bundle\") pod \"d5833467-cda7-456a-9f81-cd6b6b32a205\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.820063 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-util\") pod \"d5833467-cda7-456a-9f81-cd6b6b32a205\" (UID: \"d5833467-cda7-456a-9f81-cd6b6b32a205\") " Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.820870 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-bundle" (OuterVolumeSpecName: "bundle") pod "d5833467-cda7-456a-9f81-cd6b6b32a205" (UID: "d5833467-cda7-456a-9f81-cd6b6b32a205"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.826193 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5833467-cda7-456a-9f81-cd6b6b32a205-kube-api-access-pxz2x" (OuterVolumeSpecName: "kube-api-access-pxz2x") pod "d5833467-cda7-456a-9f81-cd6b6b32a205" (UID: "d5833467-cda7-456a-9f81-cd6b6b32a205"). InnerVolumeSpecName "kube-api-access-pxz2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.839161 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-util" (OuterVolumeSpecName: "util") pod "d5833467-cda7-456a-9f81-cd6b6b32a205" (UID: "d5833467-cda7-456a-9f81-cd6b6b32a205"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.922149 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxz2x\" (UniqueName: \"kubernetes.io/projected/d5833467-cda7-456a-9f81-cd6b6b32a205-kube-api-access-pxz2x\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.922187 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:04 crc kubenswrapper[4796]: I0930 16:25:04.922196 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d5833467-cda7-456a-9f81-cd6b6b32a205-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:05 crc kubenswrapper[4796]: I0930 16:25:05.542319 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" event={"ID":"d5833467-cda7-456a-9f81-cd6b6b32a205","Type":"ContainerDied","Data":"c2bb98a7513e6c772262c3cd47a58ebbefd6e760c8e7cff56cecdbb0efdf097c"} Sep 30 16:25:05 crc kubenswrapper[4796]: I0930 16:25:05.542359 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2bb98a7513e6c772262c3cd47a58ebbefd6e760c8e7cff56cecdbb0efdf097c" Sep 30 16:25:05 crc kubenswrapper[4796]: I0930 16:25:05.542445 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj" Sep 30 16:25:05 crc kubenswrapper[4796]: I0930 16:25:05.596014 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:25:05 crc kubenswrapper[4796]: I0930 16:25:05.596093 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:25:05 crc kubenswrapper[4796]: I0930 16:25:05.683112 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.224354 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx"] Sep 30 16:25:11 crc kubenswrapper[4796]: E0930 16:25:11.225147 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="pull" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225164 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="pull" Sep 30 16:25:11 crc kubenswrapper[4796]: E0930 16:25:11.225177 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="extract-utilities" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225184 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="extract-utilities" Sep 30 16:25:11 crc kubenswrapper[4796]: E0930 16:25:11.225192 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="extract" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225199 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="extract" Sep 30 16:25:11 crc kubenswrapper[4796]: E0930 16:25:11.225215 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="util" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225221 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="util" Sep 30 16:25:11 crc kubenswrapper[4796]: E0930 16:25:11.225233 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="registry-server" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225241 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="registry-server" Sep 30 16:25:11 crc kubenswrapper[4796]: E0930 16:25:11.225253 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="extract-content" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225259 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="extract-content" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225394 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5833467-cda7-456a-9f81-cd6b6b32a205" containerName="extract" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.225407 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a77ad7-bd47-46b0-b464-b96769d499f1" containerName="registry-server" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.226235 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.230234 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-pn49j" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.263277 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx"] Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.318060 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lshs7\" (UniqueName: \"kubernetes.io/projected/107cc98e-9b37-4a5f-99f3-73350360e2ab-kube-api-access-lshs7\") pod \"openstack-operator-controller-operator-7ff58777bc-cfdvx\" (UID: \"107cc98e-9b37-4a5f-99f3-73350360e2ab\") " pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.419956 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lshs7\" (UniqueName: \"kubernetes.io/projected/107cc98e-9b37-4a5f-99f3-73350360e2ab-kube-api-access-lshs7\") pod \"openstack-operator-controller-operator-7ff58777bc-cfdvx\" (UID: \"107cc98e-9b37-4a5f-99f3-73350360e2ab\") " pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.442106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lshs7\" (UniqueName: \"kubernetes.io/projected/107cc98e-9b37-4a5f-99f3-73350360e2ab-kube-api-access-lshs7\") pod \"openstack-operator-controller-operator-7ff58777bc-cfdvx\" (UID: \"107cc98e-9b37-4a5f-99f3-73350360e2ab\") " pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.547550 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:11 crc kubenswrapper[4796]: I0930 16:25:11.876548 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx"] Sep 30 16:25:11 crc kubenswrapper[4796]: W0930 16:25:11.886785 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod107cc98e_9b37_4a5f_99f3_73350360e2ab.slice/crio-14d5b97cf6e87aa894018f0a1e125a5ce8d60644843ab3316f58d41ae4296614 WatchSource:0}: Error finding container 14d5b97cf6e87aa894018f0a1e125a5ce8d60644843ab3316f58d41ae4296614: Status 404 returned error can't find the container with id 14d5b97cf6e87aa894018f0a1e125a5ce8d60644843ab3316f58d41ae4296614 Sep 30 16:25:12 crc kubenswrapper[4796]: I0930 16:25:12.609597 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" event={"ID":"107cc98e-9b37-4a5f-99f3-73350360e2ab","Type":"ContainerStarted","Data":"14d5b97cf6e87aa894018f0a1e125a5ce8d60644843ab3316f58d41ae4296614"} Sep 30 16:25:15 crc kubenswrapper[4796]: I0930 16:25:15.643289 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xln5d" Sep 30 16:25:16 crc kubenswrapper[4796]: I0930 16:25:16.653855 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" event={"ID":"107cc98e-9b37-4a5f-99f3-73350360e2ab","Type":"ContainerStarted","Data":"be60039911a10e2efe57a23c2b7f3f2595946e0a1b833f63f76e42e1616c4366"} Sep 30 16:25:17 crc kubenswrapper[4796]: I0930 16:25:17.711570 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xln5d"] Sep 30 16:25:18 crc kubenswrapper[4796]: I0930 16:25:18.057075 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gx5s8"] Sep 30 16:25:18 crc kubenswrapper[4796]: I0930 16:25:18.057467 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gx5s8" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="registry-server" containerID="cri-o://dc436f64dc0c06b0b89f0063900dd49204a4cbaafbaf5682372540e1696b08e3" gracePeriod=2 Sep 30 16:25:18 crc kubenswrapper[4796]: I0930 16:25:18.683217 4796 generic.go:334] "Generic (PLEG): container finished" podID="cbe44b77-d340-4354-b289-f6adb481aa27" containerID="dc436f64dc0c06b0b89f0063900dd49204a4cbaafbaf5682372540e1696b08e3" exitCode=0 Sep 30 16:25:18 crc kubenswrapper[4796]: I0930 16:25:18.683256 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx5s8" event={"ID":"cbe44b77-d340-4354-b289-f6adb481aa27","Type":"ContainerDied","Data":"dc436f64dc0c06b0b89f0063900dd49204a4cbaafbaf5682372540e1696b08e3"} Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.144816 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.226806 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-catalog-content\") pod \"cbe44b77-d340-4354-b289-f6adb481aa27\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.226875 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5nbz\" (UniqueName: \"kubernetes.io/projected/cbe44b77-d340-4354-b289-f6adb481aa27-kube-api-access-t5nbz\") pod \"cbe44b77-d340-4354-b289-f6adb481aa27\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.226924 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-utilities\") pod \"cbe44b77-d340-4354-b289-f6adb481aa27\" (UID: \"cbe44b77-d340-4354-b289-f6adb481aa27\") " Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.228139 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-utilities" (OuterVolumeSpecName: "utilities") pod "cbe44b77-d340-4354-b289-f6adb481aa27" (UID: "cbe44b77-d340-4354-b289-f6adb481aa27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.233731 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe44b77-d340-4354-b289-f6adb481aa27-kube-api-access-t5nbz" (OuterVolumeSpecName: "kube-api-access-t5nbz") pod "cbe44b77-d340-4354-b289-f6adb481aa27" (UID: "cbe44b77-d340-4354-b289-f6adb481aa27"). InnerVolumeSpecName "kube-api-access-t5nbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.283428 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbe44b77-d340-4354-b289-f6adb481aa27" (UID: "cbe44b77-d340-4354-b289-f6adb481aa27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.328672 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.328729 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe44b77-d340-4354-b289-f6adb481aa27-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.328750 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5nbz\" (UniqueName: \"kubernetes.io/projected/cbe44b77-d340-4354-b289-f6adb481aa27-kube-api-access-t5nbz\") on node \"crc\" DevicePath \"\"" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.697401 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" event={"ID":"107cc98e-9b37-4a5f-99f3-73350360e2ab","Type":"ContainerStarted","Data":"e213aa44d2eda0964c8397e01201feb0c5914e96fdb083cdc138fb27c48f57a7"} Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.699289 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.702903 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx5s8" event={"ID":"cbe44b77-d340-4354-b289-f6adb481aa27","Type":"ContainerDied","Data":"d16410c31923bc0bc25bca57102353f0e3fc4db8655ff933a32af79b692c7c1c"} Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.702975 4796 scope.go:117] "RemoveContainer" containerID="dc436f64dc0c06b0b89f0063900dd49204a4cbaafbaf5682372540e1696b08e3" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.703038 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx5s8" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.735817 4796 scope.go:117] "RemoveContainer" containerID="93456527f859d651ac332c836f05d1ffe6d42cc8bfe47102ff11d0ce76e5d272" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.775113 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" podStartSLOduration=1.7741764359999999 podStartE2EDuration="8.775058922s" podCreationTimestamp="2025-09-30 16:25:11 +0000 UTC" firstStartedPulling="2025-09-30 16:25:11.890258457 +0000 UTC m=+803.903536984" lastFinishedPulling="2025-09-30 16:25:18.891140943 +0000 UTC m=+810.904419470" observedRunningTime="2025-09-30 16:25:19.744550196 +0000 UTC m=+811.757828743" watchObservedRunningTime="2025-09-30 16:25:19.775058922 +0000 UTC m=+811.788337469" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.784062 4796 scope.go:117] "RemoveContainer" containerID="6de655918e681cbd1a0dae7542884eb1d3d332e1580d6f0e623c662a7a96e33a" Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.785658 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gx5s8"] Sep 30 16:25:19 crc kubenswrapper[4796]: I0930 16:25:19.792141 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gx5s8"] Sep 30 16:25:20 crc kubenswrapper[4796]: I0930 16:25:20.742839 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" path="/var/lib/kubelet/pods/cbe44b77-d340-4354-b289-f6adb481aa27/volumes" Sep 30 16:25:21 crc kubenswrapper[4796]: I0930 16:25:21.550597 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7ff58777bc-cfdvx" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.745045 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6"] Sep 30 16:25:49 crc kubenswrapper[4796]: E0930 16:25:49.746176 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="registry-server" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.746194 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="registry-server" Sep 30 16:25:49 crc kubenswrapper[4796]: E0930 16:25:49.746209 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="extract-content" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.746217 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="extract-content" Sep 30 16:25:49 crc kubenswrapper[4796]: E0930 16:25:49.746232 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="extract-utilities" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.746243 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="extract-utilities" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.746374 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe44b77-d340-4354-b289-f6adb481aa27" containerName="registry-server" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.747185 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.749929 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wv6kg" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.754089 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.755490 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.762930 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ggsqt" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.777106 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.783325 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.784442 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.787227 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6s9w4" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.791239 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.806171 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.814661 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.815650 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.818975 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ndzwp" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.819388 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.826122 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.827630 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.829390 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-bb6l8" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.834196 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.835208 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.837063 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-vfzhl" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.843229 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.855581 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.876200 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-97m97"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.878029 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.894686 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.895187 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-q6pkz" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.897006 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.898095 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.901701 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-77vv4" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.902566 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf8kp\" (UniqueName: \"kubernetes.io/projected/cf66eb4e-6595-40ea-b64a-b3d40e44dec9-kube-api-access-wf8kp\") pod \"glance-operator-controller-manager-84958c4d49-nvqft\" (UID: \"cf66eb4e-6595-40ea-b64a-b3d40e44dec9\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.902603 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2pw2\" (UniqueName: \"kubernetes.io/projected/8fba4c88-186e-4a06-8374-97e696008fcd-kube-api-access-f2pw2\") pod \"heat-operator-controller-manager-5d889d78cf-4dsbl\" (UID: \"8fba4c88-186e-4a06-8374-97e696008fcd\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.902685 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzs2k\" (UniqueName: \"kubernetes.io/projected/24427541-c81d-4542-b20e-bdcf3a4f0e4c-kube-api-access-tzs2k\") pod \"designate-operator-controller-manager-84f4f7b77b-44kvk\" (UID: \"24427541-c81d-4542-b20e-bdcf3a4f0e4c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.902705 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chk5b\" (UniqueName: \"kubernetes.io/projected/a6b170fa-6665-405f-aaa3-042e9705ed1a-kube-api-access-chk5b\") pod \"barbican-operator-controller-manager-6ff8b75857-dlpc6\" (UID: \"a6b170fa-6665-405f-aaa3-042e9705ed1a\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.902925 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzvlg\" (UniqueName: \"kubernetes.io/projected/52af31b2-ed79-46c4-840b-c2453d43a88c-kube-api-access-gzvlg\") pod \"cinder-operator-controller-manager-644bddb6d8-nmqln\" (UID: \"52af31b2-ed79-46c4-840b-c2453d43a88c\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.921915 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-97m97"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.943121 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.944208 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.947382 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-twgmh" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.951868 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.953216 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.956273 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-w5jvp" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.956384 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.957462 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.963356 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.972409 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qt6hj" Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.982422 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk"] Sep 30 16:25:49 crc kubenswrapper[4796]: I0930 16:25:49.990435 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.002772 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005743 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs9gx\" (UniqueName: \"kubernetes.io/projected/8ef65c41-fdac-4065-a568-2bd9f1176adc-kube-api-access-zs9gx\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005795 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzs2k\" (UniqueName: \"kubernetes.io/projected/24427541-c81d-4542-b20e-bdcf3a4f0e4c-kube-api-access-tzs2k\") pod \"designate-operator-controller-manager-84f4f7b77b-44kvk\" (UID: \"24427541-c81d-4542-b20e-bdcf3a4f0e4c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chk5b\" (UniqueName: \"kubernetes.io/projected/a6b170fa-6665-405f-aaa3-042e9705ed1a-kube-api-access-chk5b\") pod \"barbican-operator-controller-manager-6ff8b75857-dlpc6\" (UID: \"a6b170fa-6665-405f-aaa3-042e9705ed1a\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005838 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbwkm\" (UniqueName: \"kubernetes.io/projected/70f8a96e-36a5-4824-a041-72b7ccfbe064-kube-api-access-fbwkm\") pod \"horizon-operator-controller-manager-9f4696d94-jxwnq\" (UID: \"70f8a96e-36a5-4824-a041-72b7ccfbe064\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005906 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8t7z\" (UniqueName: \"kubernetes.io/projected/36b7f470-4ef5-431f-a756-04c40e1afc90-kube-api-access-x8t7z\") pod \"ironic-operator-controller-manager-7975b88857-j6vsj\" (UID: \"36b7f470-4ef5-431f-a756-04c40e1afc90\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005930 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzvlg\" (UniqueName: \"kubernetes.io/projected/52af31b2-ed79-46c4-840b-c2453d43a88c-kube-api-access-gzvlg\") pod \"cinder-operator-controller-manager-644bddb6d8-nmqln\" (UID: \"52af31b2-ed79-46c4-840b-c2453d43a88c\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005955 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8ef65c41-fdac-4065-a568-2bd9f1176adc-cert\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.005973 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2pw2\" (UniqueName: \"kubernetes.io/projected/8fba4c88-186e-4a06-8374-97e696008fcd-kube-api-access-f2pw2\") pod \"heat-operator-controller-manager-5d889d78cf-4dsbl\" (UID: \"8fba4c88-186e-4a06-8374-97e696008fcd\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.006004 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf8kp\" (UniqueName: \"kubernetes.io/projected/cf66eb4e-6595-40ea-b64a-b3d40e44dec9-kube-api-access-wf8kp\") pod \"glance-operator-controller-manager-84958c4d49-nvqft\" (UID: \"cf66eb4e-6595-40ea-b64a-b3d40e44dec9\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.017824 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.023764 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.027697 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2rr62" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.030019 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-jj598"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.057628 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.072419 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-jhg4d" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.089858 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf8kp\" (UniqueName: \"kubernetes.io/projected/cf66eb4e-6595-40ea-b64a-b3d40e44dec9-kube-api-access-wf8kp\") pod \"glance-operator-controller-manager-84958c4d49-nvqft\" (UID: \"cf66eb4e-6595-40ea-b64a-b3d40e44dec9\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.082259 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2pw2\" (UniqueName: \"kubernetes.io/projected/8fba4c88-186e-4a06-8374-97e696008fcd-kube-api-access-f2pw2\") pod \"heat-operator-controller-manager-5d889d78cf-4dsbl\" (UID: \"8fba4c88-186e-4a06-8374-97e696008fcd\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129719 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbg2\" (UniqueName: \"kubernetes.io/projected/16086739-bea6-4c47-85d8-9a10497c9373-kube-api-access-blbg2\") pod \"mariadb-operator-controller-manager-88c7-sb9j9\" (UID: \"16086739-bea6-4c47-85d8-9a10497c9373\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129759 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r6xg\" (UniqueName: \"kubernetes.io/projected/ed0b77fc-1f2a-4a53-9983-ba64a6831569-kube-api-access-4r6xg\") pod \"neutron-operator-controller-manager-64d7b59854-mttjp\" (UID: \"ed0b77fc-1f2a-4a53-9983-ba64a6831569\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129802 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8t7z\" (UniqueName: \"kubernetes.io/projected/36b7f470-4ef5-431f-a756-04c40e1afc90-kube-api-access-x8t7z\") pod \"ironic-operator-controller-manager-7975b88857-j6vsj\" (UID: \"36b7f470-4ef5-431f-a756-04c40e1afc90\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129822 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdslp\" (UniqueName: \"kubernetes.io/projected/e50b54b7-c7c3-4614-bd2e-3b90b68e129f-kube-api-access-kdslp\") pod \"keystone-operator-controller-manager-5bd55b4bff-fqxdk\" (UID: \"e50b54b7-c7c3-4614-bd2e-3b90b68e129f\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129847 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf8nw\" (UniqueName: \"kubernetes.io/projected/5a8c84fa-799a-4f47-81c1-7445702e0f23-kube-api-access-sf8nw\") pod \"manila-operator-controller-manager-6d68dbc695-j8d9f\" (UID: \"5a8c84fa-799a-4f47-81c1-7445702e0f23\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8ef65c41-fdac-4065-a568-2bd9f1176adc-cert\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129901 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs9gx\" (UniqueName: \"kubernetes.io/projected/8ef65c41-fdac-4065-a568-2bd9f1176adc-kube-api-access-zs9gx\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.129929 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbwkm\" (UniqueName: \"kubernetes.io/projected/70f8a96e-36a5-4824-a041-72b7ccfbe064-kube-api-access-fbwkm\") pod \"horizon-operator-controller-manager-9f4696d94-jxwnq\" (UID: \"70f8a96e-36a5-4824-a041-72b7ccfbe064\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:25:50 crc kubenswrapper[4796]: E0930 16:25:50.130257 4796 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Sep 30 16:25:50 crc kubenswrapper[4796]: E0930 16:25:50.130303 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef65c41-fdac-4065-a568-2bd9f1176adc-cert podName:8ef65c41-fdac-4065-a568-2bd9f1176adc nodeName:}" failed. No retries permitted until 2025-09-30 16:25:50.630288704 +0000 UTC m=+842.643567221 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8ef65c41-fdac-4065-a568-2bd9f1176adc-cert") pod "infra-operator-controller-manager-7d857cc749-97m97" (UID: "8ef65c41-fdac-4065-a568-2bd9f1176adc") : secret "infra-operator-webhook-server-cert" not found Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.134274 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzvlg\" (UniqueName: \"kubernetes.io/projected/52af31b2-ed79-46c4-840b-c2453d43a88c-kube-api-access-gzvlg\") pod \"cinder-operator-controller-manager-644bddb6d8-nmqln\" (UID: \"52af31b2-ed79-46c4-840b-c2453d43a88c\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.141567 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chk5b\" (UniqueName: \"kubernetes.io/projected/a6b170fa-6665-405f-aaa3-042e9705ed1a-kube-api-access-chk5b\") pod \"barbican-operator-controller-manager-6ff8b75857-dlpc6\" (UID: \"a6b170fa-6665-405f-aaa3-042e9705ed1a\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.143444 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.150893 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzs2k\" (UniqueName: \"kubernetes.io/projected/24427541-c81d-4542-b20e-bdcf3a4f0e4c-kube-api-access-tzs2k\") pod \"designate-operator-controller-manager-84f4f7b77b-44kvk\" (UID: \"24427541-c81d-4542-b20e-bdcf3a4f0e4c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.156873 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8t7z\" (UniqueName: \"kubernetes.io/projected/36b7f470-4ef5-431f-a756-04c40e1afc90-kube-api-access-x8t7z\") pod \"ironic-operator-controller-manager-7975b88857-j6vsj\" (UID: \"36b7f470-4ef5-431f-a756-04c40e1afc90\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.159383 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.160422 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs9gx\" (UniqueName: \"kubernetes.io/projected/8ef65c41-fdac-4065-a568-2bd9f1176adc-kube-api-access-zs9gx\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.160522 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbwkm\" (UniqueName: \"kubernetes.io/projected/70f8a96e-36a5-4824-a041-72b7ccfbe064-kube-api-access-fbwkm\") pod \"horizon-operator-controller-manager-9f4696d94-jxwnq\" (UID: \"70f8a96e-36a5-4824-a041-72b7ccfbe064\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.161655 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.174183 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.177132 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-jj598"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.189444 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.191138 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.193026 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.193302 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-kknxp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.197734 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.198899 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.200809 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-n2ttk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.213262 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.220182 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.234204 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.234846 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbg2\" (UniqueName: \"kubernetes.io/projected/16086739-bea6-4c47-85d8-9a10497c9373-kube-api-access-blbg2\") pod \"mariadb-operator-controller-manager-88c7-sb9j9\" (UID: \"16086739-bea6-4c47-85d8-9a10497c9373\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.234884 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r6xg\" (UniqueName: \"kubernetes.io/projected/ed0b77fc-1f2a-4a53-9983-ba64a6831569-kube-api-access-4r6xg\") pod \"neutron-operator-controller-manager-64d7b59854-mttjp\" (UID: \"ed0b77fc-1f2a-4a53-9983-ba64a6831569\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.234932 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zld5h\" (UniqueName: \"kubernetes.io/projected/4131038b-1bbe-410a-9c7b-58216c527106-kube-api-access-zld5h\") pod \"nova-operator-controller-manager-c7c776c96-jj598\" (UID: \"4131038b-1bbe-410a-9c7b-58216c527106\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.234952 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdslp\" (UniqueName: \"kubernetes.io/projected/e50b54b7-c7c3-4614-bd2e-3b90b68e129f-kube-api-access-kdslp\") pod \"keystone-operator-controller-manager-5bd55b4bff-fqxdk\" (UID: \"e50b54b7-c7c3-4614-bd2e-3b90b68e129f\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.235003 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf8nw\" (UniqueName: \"kubernetes.io/projected/5a8c84fa-799a-4f47-81c1-7445702e0f23-kube-api-access-sf8nw\") pod \"manila-operator-controller-manager-6d68dbc695-j8d9f\" (UID: \"5a8c84fa-799a-4f47-81c1-7445702e0f23\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.235412 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.242194 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.245124 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-cs48r" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.261671 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdslp\" (UniqueName: \"kubernetes.io/projected/e50b54b7-c7c3-4614-bd2e-3b90b68e129f-kube-api-access-kdslp\") pod \"keystone-operator-controller-manager-5bd55b4bff-fqxdk\" (UID: \"e50b54b7-c7c3-4614-bd2e-3b90b68e129f\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.266872 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.283606 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r6xg\" (UniqueName: \"kubernetes.io/projected/ed0b77fc-1f2a-4a53-9983-ba64a6831569-kube-api-access-4r6xg\") pod \"neutron-operator-controller-manager-64d7b59854-mttjp\" (UID: \"ed0b77fc-1f2a-4a53-9983-ba64a6831569\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.283889 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.286457 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbg2\" (UniqueName: \"kubernetes.io/projected/16086739-bea6-4c47-85d8-9a10497c9373-kube-api-access-blbg2\") pod \"mariadb-operator-controller-manager-88c7-sb9j9\" (UID: \"16086739-bea6-4c47-85d8-9a10497c9373\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.287875 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf8nw\" (UniqueName: \"kubernetes.io/projected/5a8c84fa-799a-4f47-81c1-7445702e0f23-kube-api-access-sf8nw\") pod \"manila-operator-controller-manager-6d68dbc695-j8d9f\" (UID: \"5a8c84fa-799a-4f47-81c1-7445702e0f23\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.292932 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.294469 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.305021 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-hh67b" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.310219 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.314448 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.316489 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4xrvq" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.326355 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.340392 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df653968-5944-4e8d-9e82-7e19a539d997-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.340460 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbz5\" (UniqueName: \"kubernetes.io/projected/5cadf327-7ef7-439e-b692-5eef55c1b666-kube-api-access-wgbz5\") pod \"ovn-operator-controller-manager-9976ff44c-jqbdb\" (UID: \"5cadf327-7ef7-439e-b692-5eef55c1b666\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.340540 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zld5h\" (UniqueName: \"kubernetes.io/projected/4131038b-1bbe-410a-9c7b-58216c527106-kube-api-access-zld5h\") pod \"nova-operator-controller-manager-c7c776c96-jj598\" (UID: \"4131038b-1bbe-410a-9c7b-58216c527106\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.340560 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nghf\" (UniqueName: \"kubernetes.io/projected/df653968-5944-4e8d-9e82-7e19a539d997-kube-api-access-4nghf\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.340582 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgm5f\" (UniqueName: \"kubernetes.io/projected/d1555f09-a29d-45bf-8bdf-5a3212d1ce0a-kube-api-access-bgm5f\") pod \"octavia-operator-controller-manager-76fcc6dc7c-q2knx\" (UID: \"d1555f09-a29d-45bf-8bdf-5a3212d1ce0a\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.355535 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.363151 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.375560 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.375726 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.377170 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.377930 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.381532 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-psjfd" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.384404 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zld5h\" (UniqueName: \"kubernetes.io/projected/4131038b-1bbe-410a-9c7b-58216c527106-kube-api-access-zld5h\") pod \"nova-operator-controller-manager-c7c776c96-jj598\" (UID: \"4131038b-1bbe-410a-9c7b-58216c527106\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.384509 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.388355 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.404821 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.420870 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-dntbx"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.421967 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.425058 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-jx4m5" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.428516 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-dntbx"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.443871 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df653968-5944-4e8d-9e82-7e19a539d997-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.443931 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbz5\" (UniqueName: \"kubernetes.io/projected/5cadf327-7ef7-439e-b692-5eef55c1b666-kube-api-access-wgbz5\") pod \"ovn-operator-controller-manager-9976ff44c-jqbdb\" (UID: \"5cadf327-7ef7-439e-b692-5eef55c1b666\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.444011 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nghf\" (UniqueName: \"kubernetes.io/projected/df653968-5944-4e8d-9e82-7e19a539d997-kube-api-access-4nghf\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.444040 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgm5f\" (UniqueName: \"kubernetes.io/projected/d1555f09-a29d-45bf-8bdf-5a3212d1ce0a-kube-api-access-bgm5f\") pod \"octavia-operator-controller-manager-76fcc6dc7c-q2knx\" (UID: \"d1555f09-a29d-45bf-8bdf-5a3212d1ce0a\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.444077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5m8b\" (UniqueName: \"kubernetes.io/projected/ae3399ed-5451-44ba-9872-73d709faab31-kube-api-access-b5m8b\") pod \"placement-operator-controller-manager-589c58c6c-whxjg\" (UID: \"ae3399ed-5451-44ba-9872-73d709faab31\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.444134 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdp89\" (UniqueName: \"kubernetes.io/projected/cb7865b9-6ef2-4cc1-92e5-0753a28e43c1-kube-api-access-zdp89\") pod \"swift-operator-controller-manager-bc7dc7bd9-f45lz\" (UID: \"cb7865b9-6ef2-4cc1-92e5-0753a28e43c1\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:25:50 crc kubenswrapper[4796]: E0930 16:25:50.444344 4796 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 16:25:50 crc kubenswrapper[4796]: E0930 16:25:50.444397 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df653968-5944-4e8d-9e82-7e19a539d997-cert podName:df653968-5944-4e8d-9e82-7e19a539d997 nodeName:}" failed. No retries permitted until 2025-09-30 16:25:50.944379634 +0000 UTC m=+842.957658161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/df653968-5944-4e8d-9e82-7e19a539d997-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-p2jwc" (UID: "df653968-5944-4e8d-9e82-7e19a539d997") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.458461 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.459599 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.478662 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.480322 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-82jzb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.484685 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbz5\" (UniqueName: \"kubernetes.io/projected/5cadf327-7ef7-439e-b692-5eef55c1b666-kube-api-access-wgbz5\") pod \"ovn-operator-controller-manager-9976ff44c-jqbdb\" (UID: \"5cadf327-7ef7-439e-b692-5eef55c1b666\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.496109 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nghf\" (UniqueName: \"kubernetes.io/projected/df653968-5944-4e8d-9e82-7e19a539d997-kube-api-access-4nghf\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.497069 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.498309 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.539752 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.539928 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgm5f\" (UniqueName: \"kubernetes.io/projected/d1555f09-a29d-45bf-8bdf-5a3212d1ce0a-kube-api-access-bgm5f\") pod \"octavia-operator-controller-manager-76fcc6dc7c-q2knx\" (UID: \"d1555f09-a29d-45bf-8bdf-5a3212d1ce0a\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.542837 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-d62p2" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.543728 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.547563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjbv4\" (UniqueName: \"kubernetes.io/projected/530de9cb-0bdc-4773-a4ff-29b992d1942f-kube-api-access-hjbv4\") pod \"telemetry-operator-controller-manager-6db74cfd6d-l8qfg\" (UID: \"530de9cb-0bdc-4773-a4ff-29b992d1942f\") " pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.547655 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5m8b\" (UniqueName: \"kubernetes.io/projected/ae3399ed-5451-44ba-9872-73d709faab31-kube-api-access-b5m8b\") pod \"placement-operator-controller-manager-589c58c6c-whxjg\" (UID: \"ae3399ed-5451-44ba-9872-73d709faab31\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.547748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdp89\" (UniqueName: \"kubernetes.io/projected/cb7865b9-6ef2-4cc1-92e5-0753a28e43c1-kube-api-access-zdp89\") pod \"swift-operator-controller-manager-bc7dc7bd9-f45lz\" (UID: \"cb7865b9-6ef2-4cc1-92e5-0753a28e43c1\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.547815 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78gjt\" (UniqueName: \"kubernetes.io/projected/2a580ec9-b0df-41bb-a272-cc4076cacdba-kube-api-access-78gjt\") pod \"watcher-operator-controller-manager-76669f99c-j4ntp\" (UID: \"2a580ec9-b0df-41bb-a272-cc4076cacdba\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.547880 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56wv5\" (UniqueName: \"kubernetes.io/projected/40fc5188-af1b-454f-b7f4-6c1ae1c4cf71-kube-api-access-56wv5\") pod \"test-operator-controller-manager-f66b554c6-dntbx\" (UID: \"40fc5188-af1b-454f-b7f4-6c1ae1c4cf71\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.565885 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.586761 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.593183 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdp89\" (UniqueName: \"kubernetes.io/projected/cb7865b9-6ef2-4cc1-92e5-0753a28e43c1-kube-api-access-zdp89\") pod \"swift-operator-controller-manager-bc7dc7bd9-f45lz\" (UID: \"cb7865b9-6ef2-4cc1-92e5-0753a28e43c1\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.594632 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5m8b\" (UniqueName: \"kubernetes.io/projected/ae3399ed-5451-44ba-9872-73d709faab31-kube-api-access-b5m8b\") pod \"placement-operator-controller-manager-589c58c6c-whxjg\" (UID: \"ae3399ed-5451-44ba-9872-73d709faab31\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.596686 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.597652 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.600829 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-7sht8" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.612434 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.616237 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m"] Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.630771 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.649530 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xrdg\" (UniqueName: \"kubernetes.io/projected/d9b3d318-c9a7-412a-a926-6f840dd14df6-kube-api-access-7xrdg\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.649617 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78gjt\" (UniqueName: \"kubernetes.io/projected/2a580ec9-b0df-41bb-a272-cc4076cacdba-kube-api-access-78gjt\") pod \"watcher-operator-controller-manager-76669f99c-j4ntp\" (UID: \"2a580ec9-b0df-41bb-a272-cc4076cacdba\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.649665 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d9b3d318-c9a7-412a-a926-6f840dd14df6-cert\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.649686 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56wv5\" (UniqueName: \"kubernetes.io/projected/40fc5188-af1b-454f-b7f4-6c1ae1c4cf71-kube-api-access-56wv5\") pod \"test-operator-controller-manager-f66b554c6-dntbx\" (UID: \"40fc5188-af1b-454f-b7f4-6c1ae1c4cf71\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.649711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjbv4\" (UniqueName: \"kubernetes.io/projected/530de9cb-0bdc-4773-a4ff-29b992d1942f-kube-api-access-hjbv4\") pod \"telemetry-operator-controller-manager-6db74cfd6d-l8qfg\" (UID: \"530de9cb-0bdc-4773-a4ff-29b992d1942f\") " pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.649744 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8ef65c41-fdac-4065-a568-2bd9f1176adc-cert\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.656357 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.674680 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.677209 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8ef65c41-fdac-4065-a568-2bd9f1176adc-cert\") pod \"infra-operator-controller-manager-7d857cc749-97m97\" (UID: \"8ef65c41-fdac-4065-a568-2bd9f1176adc\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.686172 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78gjt\" (UniqueName: \"kubernetes.io/projected/2a580ec9-b0df-41bb-a272-cc4076cacdba-kube-api-access-78gjt\") pod \"watcher-operator-controller-manager-76669f99c-j4ntp\" (UID: \"2a580ec9-b0df-41bb-a272-cc4076cacdba\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.691265 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjbv4\" (UniqueName: \"kubernetes.io/projected/530de9cb-0bdc-4773-a4ff-29b992d1942f-kube-api-access-hjbv4\") pod \"telemetry-operator-controller-manager-6db74cfd6d-l8qfg\" (UID: \"530de9cb-0bdc-4773-a4ff-29b992d1942f\") " pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.696568 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56wv5\" (UniqueName: \"kubernetes.io/projected/40fc5188-af1b-454f-b7f4-6c1ae1c4cf71-kube-api-access-56wv5\") pod \"test-operator-controller-manager-f66b554c6-dntbx\" (UID: \"40fc5188-af1b-454f-b7f4-6c1ae1c4cf71\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.750557 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xrdg\" (UniqueName: \"kubernetes.io/projected/d9b3d318-c9a7-412a-a926-6f840dd14df6-kube-api-access-7xrdg\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.750689 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dvgk\" (UniqueName: \"kubernetes.io/projected/d70c88cd-5ff3-49d9-b581-54860881ea39-kube-api-access-9dvgk\") pod \"rabbitmq-cluster-operator-manager-79d8469568-dwv6m\" (UID: \"d70c88cd-5ff3-49d9-b581-54860881ea39\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.750718 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d9b3d318-c9a7-412a-a926-6f840dd14df6-cert\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:50 crc kubenswrapper[4796]: E0930 16:25:50.750866 4796 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Sep 30 16:25:50 crc kubenswrapper[4796]: E0930 16:25:50.750925 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b3d318-c9a7-412a-a926-6f840dd14df6-cert podName:d9b3d318-c9a7-412a-a926-6f840dd14df6 nodeName:}" failed. No retries permitted until 2025-09-30 16:25:51.250907927 +0000 UTC m=+843.264186454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d9b3d318-c9a7-412a-a926-6f840dd14df6-cert") pod "openstack-operator-controller-manager-55476bd9c7-gvldb" (UID: "d9b3d318-c9a7-412a-a926-6f840dd14df6") : secret "webhook-server-cert" not found Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.774822 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xrdg\" (UniqueName: \"kubernetes.io/projected/d9b3d318-c9a7-412a-a926-6f840dd14df6-kube-api-access-7xrdg\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.825093 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.851523 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dvgk\" (UniqueName: \"kubernetes.io/projected/d70c88cd-5ff3-49d9-b581-54860881ea39-kube-api-access-9dvgk\") pod \"rabbitmq-cluster-operator-manager-79d8469568-dwv6m\" (UID: \"d70c88cd-5ff3-49d9-b581-54860881ea39\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.869630 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dvgk\" (UniqueName: \"kubernetes.io/projected/d70c88cd-5ff3-49d9-b581-54860881ea39-kube-api-access-9dvgk\") pod \"rabbitmq-cluster-operator-manager-79d8469568-dwv6m\" (UID: \"d70c88cd-5ff3-49d9-b581-54860881ea39\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.899610 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.925548 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.946631 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.952499 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df653968-5944-4e8d-9e82-7e19a539d997-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:50 crc kubenswrapper[4796]: I0930 16:25:50.961243 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df653968-5944-4e8d-9e82-7e19a539d997-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p2jwc\" (UID: \"df653968-5944-4e8d-9e82-7e19a539d997\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.025273 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.197487 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.256833 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d9b3d318-c9a7-412a-a926-6f840dd14df6-cert\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.266472 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d9b3d318-c9a7-412a-a926-6f840dd14df6-cert\") pod \"openstack-operator-controller-manager-55476bd9c7-gvldb\" (UID: \"d9b3d318-c9a7-412a-a926-6f840dd14df6\") " pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.298234 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.390935 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.401145 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq"] Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.411528 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf66eb4e_6595_40ea_b64a_b3d40e44dec9.slice/crio-534abd31b6c6d8f2bc27023cbcc08d977c1b58c94386154860340f9dcb0de628 WatchSource:0}: Error finding container 534abd31b6c6d8f2bc27023cbcc08d977c1b58c94386154860340f9dcb0de628: Status 404 returned error can't find the container with id 534abd31b6c6d8f2bc27023cbcc08d977c1b58c94386154860340f9dcb0de628 Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.413002 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70f8a96e_36a5_4824_a041_72b7ccfbe064.slice/crio-c2ca13e7749632d364bd4cfeba0b11617ee46e1d82fdd1f077d0f09592f0be91 WatchSource:0}: Error finding container c2ca13e7749632d364bd4cfeba0b11617ee46e1d82fdd1f077d0f09592f0be91: Status 404 returned error can't find the container with id c2ca13e7749632d364bd4cfeba0b11617ee46e1d82fdd1f077d0f09592f0be91 Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.421104 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.737721 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp"] Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.742320 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded0b77fc_1f2a_4a53_9983_ba64a6831569.slice/crio-34a3eda20e552188ad06f7bba170c8bfebb0b27622c297c3d14da6eebaae9024 WatchSource:0}: Error finding container 34a3eda20e552188ad06f7bba170c8bfebb0b27622c297c3d14da6eebaae9024: Status 404 returned error can't find the container with id 34a3eda20e552188ad06f7bba170c8bfebb0b27622c297c3d14da6eebaae9024 Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.763022 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk"] Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.767674 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode50b54b7_c7c3_4614_bd2e_3b90b68e129f.slice/crio-6ca6675451cd7925da0b764e82c3efd76eca71ab1fd456bd5c2d5e93294132a5 WatchSource:0}: Error finding container 6ca6675451cd7925da0b764e82c3efd76eca71ab1fd456bd5c2d5e93294132a5: Status 404 returned error can't find the container with id 6ca6675451cd7925da0b764e82c3efd76eca71ab1fd456bd5c2d5e93294132a5 Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.770935 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.781404 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl"] Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.788708 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24427541_c81d_4542_b20e_bdcf3a4f0e4c.slice/crio-ad3ddd5dadb5b1564c12a706fea23e526a3c0653d41669d64042d174ff123a5c WatchSource:0}: Error finding container ad3ddd5dadb5b1564c12a706fea23e526a3c0653d41669d64042d174ff123a5c: Status 404 returned error can't find the container with id ad3ddd5dadb5b1564c12a706fea23e526a3c0653d41669d64042d174ff123a5c Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.791117 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fba4c88_186e_4a06_8374_97e696008fcd.slice/crio-d37803ee2f439fbea0d04734e9c7d07a9cb58a8d4bbad20a79d46256cef2fbf7 WatchSource:0}: Error finding container d37803ee2f439fbea0d04734e9c7d07a9cb58a8d4bbad20a79d46256cef2fbf7: Status 404 returned error can't find the container with id d37803ee2f439fbea0d04734e9c7d07a9cb58a8d4bbad20a79d46256cef2fbf7 Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.840662 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16086739_bea6_4c47_85d8_9a10497c9373.slice/crio-10237a66ea1a5446b244c3cabde86425299ce737160f3d3864e1a6d5c1c8f80c WatchSource:0}: Error finding container 10237a66ea1a5446b244c3cabde86425299ce737160f3d3864e1a6d5c1c8f80c: Status 404 returned error can't find the container with id 10237a66ea1a5446b244c3cabde86425299ce737160f3d3864e1a6d5c1c8f80c Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.842276 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.860476 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.867017 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-jj598"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.875603 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6"] Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.882743 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6b170fa_6665_405f_aaa3_042e9705ed1a.slice/crio-7fc602080172e14b058ccf475236ecfa92ea7a187f9799810754f3a32419d5c2 WatchSource:0}: Error finding container 7fc602080172e14b058ccf475236ecfa92ea7a187f9799810754f3a32419d5c2: Status 404 returned error can't find the container with id 7fc602080172e14b058ccf475236ecfa92ea7a187f9799810754f3a32419d5c2 Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.882950 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.897501 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.917132 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f"] Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.925664 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a8c84fa_799a_4f47_81c1_7445702e0f23.slice/crio-3be0b74d15cecaf11eb2bc07fa1a995a2f56cc7e1dcacc7dda5edba52649f0ac WatchSource:0}: Error finding container 3be0b74d15cecaf11eb2bc07fa1a995a2f56cc7e1dcacc7dda5edba52649f0ac: Status 404 returned error can't find the container with id 3be0b74d15cecaf11eb2bc07fa1a995a2f56cc7e1dcacc7dda5edba52649f0ac Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.927065 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52af31b2_ed79_46c4_840b_c2453d43a88c.slice/crio-05759235e730ffcf17ed1fdee79dadc83370a23ea3492f10460da4e68d50124c WatchSource:0}: Error finding container 05759235e730ffcf17ed1fdee79dadc83370a23ea3492f10460da4e68d50124c: Status 404 returned error can't find the container with id 05759235e730ffcf17ed1fdee79dadc83370a23ea3492f10460da4e68d50124c Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.927450 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln"] Sep 30 16:25:51 crc kubenswrapper[4796]: E0930 16:25:51.928736 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sf8nw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-j8d9f_openstack-operators(5a8c84fa-799a-4f47-81c1-7445702e0f23): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:51 crc kubenswrapper[4796]: W0930 16:25:51.947152 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cadf327_7ef7_439e_b692_5eef55c1b666.slice/crio-8a63d6c470a7407cdf1add6fa67896a73b1d821465dc252543d42ff92e6ac43a WatchSource:0}: Error finding container 8a63d6c470a7407cdf1add6fa67896a73b1d821465dc252543d42ff92e6ac43a: Status 404 returned error can't find the container with id 8a63d6c470a7407cdf1add6fa67896a73b1d821465dc252543d42ff92e6ac43a Sep 30 16:25:51 crc kubenswrapper[4796]: E0930 16:25:51.949163 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zdp89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-bc7dc7bd9-f45lz_openstack-operators(cb7865b9-6ef2-4cc1-92e5-0753a28e43c1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:51 crc kubenswrapper[4796]: E0930 16:25:51.949408 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gzvlg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-644bddb6d8-nmqln_openstack-operators(52af31b2-ed79-46c4-840b-c2453d43a88c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:51 crc kubenswrapper[4796]: E0930 16:25:51.955180 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wgbz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-jqbdb_openstack-operators(5cadf327-7ef7-439e-b692-5eef55c1b666): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.956862 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.972254 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb"] Sep 30 16:25:51 crc kubenswrapper[4796]: I0930 16:25:51.988753 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" event={"ID":"52af31b2-ed79-46c4-840b-c2453d43a88c","Type":"ContainerStarted","Data":"05759235e730ffcf17ed1fdee79dadc83370a23ea3492f10460da4e68d50124c"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.008150 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" event={"ID":"24427541-c81d-4542-b20e-bdcf3a4f0e4c","Type":"ContainerStarted","Data":"ad3ddd5dadb5b1564c12a706fea23e526a3c0653d41669d64042d174ff123a5c"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.020562 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" event={"ID":"ae3399ed-5451-44ba-9872-73d709faab31","Type":"ContainerStarted","Data":"e3e96ea4663151fc626ec1c6d9c135589d7e2a89e2500ebfbc95078b941a1c8c"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.022731 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" event={"ID":"e50b54b7-c7c3-4614-bd2e-3b90b68e129f","Type":"ContainerStarted","Data":"6ca6675451cd7925da0b764e82c3efd76eca71ab1fd456bd5c2d5e93294132a5"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.033567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" event={"ID":"d1555f09-a29d-45bf-8bdf-5a3212d1ce0a","Type":"ContainerStarted","Data":"256e552f8da9373027c57aba7e33f287a0be097e843123f0214f58c4a412eab9"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.035508 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" event={"ID":"4131038b-1bbe-410a-9c7b-58216c527106","Type":"ContainerStarted","Data":"70391ef449424d12d7243a5dd87617b998a169b3a4e767ea94da5f82c121fa75"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.037619 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" event={"ID":"a6b170fa-6665-405f-aaa3-042e9705ed1a","Type":"ContainerStarted","Data":"7fc602080172e14b058ccf475236ecfa92ea7a187f9799810754f3a32419d5c2"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.039323 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" event={"ID":"16086739-bea6-4c47-85d8-9a10497c9373","Type":"ContainerStarted","Data":"10237a66ea1a5446b244c3cabde86425299ce737160f3d3864e1a6d5c1c8f80c"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.042490 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" event={"ID":"8fba4c88-186e-4a06-8374-97e696008fcd","Type":"ContainerStarted","Data":"d37803ee2f439fbea0d04734e9c7d07a9cb58a8d4bbad20a79d46256cef2fbf7"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.044523 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" event={"ID":"cb7865b9-6ef2-4cc1-92e5-0753a28e43c1","Type":"ContainerStarted","Data":"65a60f7c47681c5012a0fbf48b545dd85b98789497de5d2489c584b5f368941d"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.050502 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" event={"ID":"5a8c84fa-799a-4f47-81c1-7445702e0f23","Type":"ContainerStarted","Data":"3be0b74d15cecaf11eb2bc07fa1a995a2f56cc7e1dcacc7dda5edba52649f0ac"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.051912 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" event={"ID":"70f8a96e-36a5-4824-a041-72b7ccfbe064","Type":"ContainerStarted","Data":"c2ca13e7749632d364bd4cfeba0b11617ee46e1d82fdd1f077d0f09592f0be91"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.054642 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" event={"ID":"36b7f470-4ef5-431f-a756-04c40e1afc90","Type":"ContainerStarted","Data":"11354886af2711f12276a97faf95c8eed5003975e8d013f8ad5940570cd3fe7f"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.057140 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" event={"ID":"cf66eb4e-6595-40ea-b64a-b3d40e44dec9","Type":"ContainerStarted","Data":"534abd31b6c6d8f2bc27023cbcc08d977c1b58c94386154860340f9dcb0de628"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.064164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" event={"ID":"ed0b77fc-1f2a-4a53-9983-ba64a6831569","Type":"ContainerStarted","Data":"34a3eda20e552188ad06f7bba170c8bfebb0b27622c297c3d14da6eebaae9024"} Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.123786 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-97m97"] Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.134259 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-dntbx"] Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.150511 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc"] Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.165426 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-56wv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-f66b554c6-dntbx_openstack-operators(40fc5188-af1b-454f-b7f4-6c1ae1c4cf71): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.169907 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.45:5001/openstack-k8s-operators/telemetry-operator:7c19824d352ab9e44515a7d4cdeb48089ae0bcf1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjbv4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6db74cfd6d-l8qfg_openstack-operators(530de9cb-0bdc-4773-a4ff-29b992d1942f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:52 crc kubenswrapper[4796]: W0930 16:25:52.176384 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf653968_5944_4e8d_9e82_7e19a539d997.slice/crio-c199add5defe6c013180063e66bc598acfb13f47ac07572b52405042ab03f822 WatchSource:0}: Error finding container c199add5defe6c013180063e66bc598acfb13f47ac07572b52405042ab03f822: Status 404 returned error can't find the container with id c199add5defe6c013180063e66bc598acfb13f47ac07572b52405042ab03f822 Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.180778 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m"] Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.194162 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg"] Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.195126 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-78gjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-76669f99c-j4ntp_openstack-operators(2a580ec9-b0df-41bb-a272-cc4076cacdba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.196850 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_LIGHTSPEED_IMAGE_URL_DEFAULT,Value:quay.io/openstack-lightspeed/rag-content:os-docs-2024.2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4nghf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6d776955-p2jwc_openstack-operators(df653968-5944-4e8d-9e82-7e19a539d997): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.197273 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:de99ad053f95f132f62b38335b2e8bf22fc28acbd441c3814764d63b63ef755f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zs9gx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-7d857cc749-97m97_openstack-operators(8ef65c41-fdac-4065-a568-2bd9f1176adc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.201881 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp"] Sep 30 16:25:52 crc kubenswrapper[4796]: I0930 16:25:52.219235 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb"] Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.388806 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" podUID="cb7865b9-6ef2-4cc1-92e5-0753a28e43c1" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.400186 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" podUID="5a8c84fa-799a-4f47-81c1-7445702e0f23" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.497914 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" podUID="40fc5188-af1b-454f-b7f4-6c1ae1c4cf71" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.498478 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" podUID="5cadf327-7ef7-439e-b692-5eef55c1b666" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.595168 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" podUID="52af31b2-ed79-46c4-840b-c2453d43a88c" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.636739 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" podUID="8ef65c41-fdac-4065-a568-2bd9f1176adc" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.707393 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" podUID="530de9cb-0bdc-4773-a4ff-29b992d1942f" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.779682 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" podUID="2a580ec9-b0df-41bb-a272-cc4076cacdba" Sep 30 16:25:52 crc kubenswrapper[4796]: E0930 16:25:52.976267 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" podUID="df653968-5944-4e8d-9e82-7e19a539d997" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.080896 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" event={"ID":"5cadf327-7ef7-439e-b692-5eef55c1b666","Type":"ContainerStarted","Data":"6a1c6f41252aa0c9441ee74abec6c7ca8be1ce1b90bad412bbd8870aa87e2f2b"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.080954 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" event={"ID":"5cadf327-7ef7-439e-b692-5eef55c1b666","Type":"ContainerStarted","Data":"8a63d6c470a7407cdf1add6fa67896a73b1d821465dc252543d42ff92e6ac43a"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.084205 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" podUID="5cadf327-7ef7-439e-b692-5eef55c1b666" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.150110 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" event={"ID":"d70c88cd-5ff3-49d9-b581-54860881ea39","Type":"ContainerStarted","Data":"23595298acd5016c6f8a5c56f7f00a99b6ca1c2b371bed99c716ffde9ec7df8e"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.153103 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" event={"ID":"2a580ec9-b0df-41bb-a272-cc4076cacdba","Type":"ContainerStarted","Data":"de18d1f6c074a14988e8670f1f3b198aec214be37cb9df7da1f3957477f52cdd"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.153132 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" event={"ID":"2a580ec9-b0df-41bb-a272-cc4076cacdba","Type":"ContainerStarted","Data":"b11532e3012920644801ac437d758ab87ccb168bb08877a8dcdad6f007acd4be"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.158125 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" podUID="2a580ec9-b0df-41bb-a272-cc4076cacdba" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.162048 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" event={"ID":"52af31b2-ed79-46c4-840b-c2453d43a88c","Type":"ContainerStarted","Data":"12c014f387fbdee776de5539c60d946a52ac45a10cb8cccf3c1e9e0be4f3968b"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.163622 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" podUID="52af31b2-ed79-46c4-840b-c2453d43a88c" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.165088 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" event={"ID":"d9b3d318-c9a7-412a-a926-6f840dd14df6","Type":"ContainerStarted","Data":"f75d882574b0d10cb39e0033dec7f98c88b964a06371742afc0cd68528c6eb29"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.165115 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" event={"ID":"d9b3d318-c9a7-412a-a926-6f840dd14df6","Type":"ContainerStarted","Data":"3be2238598c9662b114311914184bbc4ab1d5cd3dfaecf32302ed9b998ec5e48"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.165536 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.178800 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" event={"ID":"cb7865b9-6ef2-4cc1-92e5-0753a28e43c1","Type":"ContainerStarted","Data":"d2b41eea20d7741e4cd32247b7f871235b077e526de659412c72a66a59a42757"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.184728 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" event={"ID":"df653968-5944-4e8d-9e82-7e19a539d997","Type":"ContainerStarted","Data":"92c2a860d7403f474f7812c473d32dd5a9158c013483eb0e9ea4d9ce98144a03"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.184764 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" event={"ID":"df653968-5944-4e8d-9e82-7e19a539d997","Type":"ContainerStarted","Data":"c199add5defe6c013180063e66bc598acfb13f47ac07572b52405042ab03f822"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.185416 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" podUID="cb7865b9-6ef2-4cc1-92e5-0753a28e43c1" Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.185751 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" podUID="df653968-5944-4e8d-9e82-7e19a539d997" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.186574 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" event={"ID":"40fc5188-af1b-454f-b7f4-6c1ae1c4cf71","Type":"ContainerStarted","Data":"4b69a1a35791179e43a81358c80ae6d1be48cc6ea6bfca663bcb95eb121aa540"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.186598 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" event={"ID":"40fc5188-af1b-454f-b7f4-6c1ae1c4cf71","Type":"ContainerStarted","Data":"6baefe088a1d65dc5601bbbc01413468c5fa6039eb26c69003ee0bcc056f39a2"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.190577 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" podUID="40fc5188-af1b-454f-b7f4-6c1ae1c4cf71" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.193214 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" event={"ID":"8ef65c41-fdac-4065-a568-2bd9f1176adc","Type":"ContainerStarted","Data":"8fe9e7743c5e8fa382b771e0be7e43b33573f16645f73a4ee7d7148d952daaf1"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.193264 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" event={"ID":"8ef65c41-fdac-4065-a568-2bd9f1176adc","Type":"ContainerStarted","Data":"9c8f7a41498bbb417a52e25502ea7a6f57c7d0f923adf1d5c140cf9c8b93895e"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.203334 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:de99ad053f95f132f62b38335b2e8bf22fc28acbd441c3814764d63b63ef755f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" podUID="8ef65c41-fdac-4065-a568-2bd9f1176adc" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.208756 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" event={"ID":"530de9cb-0bdc-4773-a4ff-29b992d1942f","Type":"ContainerStarted","Data":"9d11280a1048a71df63c7249bb83c72a60a5516337b1ccde01b32454c46d9378"} Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.208814 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" event={"ID":"530de9cb-0bdc-4773-a4ff-29b992d1942f","Type":"ContainerStarted","Data":"20b72d9561b00e3321b40227bfc1cdfd37567a2fed2e6f3d79b2d1c046408e38"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.210467 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.45:5001/openstack-k8s-operators/telemetry-operator:7c19824d352ab9e44515a7d4cdeb48089ae0bcf1\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" podUID="530de9cb-0bdc-4773-a4ff-29b992d1942f" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.211721 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" event={"ID":"5a8c84fa-799a-4f47-81c1-7445702e0f23","Type":"ContainerStarted","Data":"5e50b4098cc03edd44042af79c4bf5b6aa36c5bc5578db5bbd455dbc266e472e"} Sep 30 16:25:53 crc kubenswrapper[4796]: E0930 16:25:53.213176 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" podUID="5a8c84fa-799a-4f47-81c1-7445702e0f23" Sep 30 16:25:53 crc kubenswrapper[4796]: I0930 16:25:53.217350 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" podStartSLOduration=3.21733439 podStartE2EDuration="3.21733439s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:25:53.213551806 +0000 UTC m=+845.226830333" watchObservedRunningTime="2025-09-30 16:25:53.21733439 +0000 UTC m=+845.230612907" Sep 30 16:25:54 crc kubenswrapper[4796]: I0930 16:25:54.228284 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" event={"ID":"d9b3d318-c9a7-412a-a926-6f840dd14df6","Type":"ContainerStarted","Data":"70795768e9156532c616f2ebefc485eb0190da66992747cba2091cc29618026f"} Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238269 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" podUID="5cadf327-7ef7-439e-b692-5eef55c1b666" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238331 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:de99ad053f95f132f62b38335b2e8bf22fc28acbd441c3814764d63b63ef755f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" podUID="8ef65c41-fdac-4065-a568-2bd9f1176adc" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238408 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.45:5001/openstack-k8s-operators/telemetry-operator:7c19824d352ab9e44515a7d4cdeb48089ae0bcf1\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" podUID="530de9cb-0bdc-4773-a4ff-29b992d1942f" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238459 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" podUID="cb7865b9-6ef2-4cc1-92e5-0753a28e43c1" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238507 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" podUID="52af31b2-ed79-46c4-840b-c2453d43a88c" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238553 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" podUID="df653968-5944-4e8d-9e82-7e19a539d997" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238613 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" podUID="5a8c84fa-799a-4f47-81c1-7445702e0f23" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.238664 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" podUID="2a580ec9-b0df-41bb-a272-cc4076cacdba" Sep 30 16:25:54 crc kubenswrapper[4796]: E0930 16:25:54.239168 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" podUID="40fc5188-af1b-454f-b7f4-6c1ae1c4cf71" Sep 30 16:26:01 crc kubenswrapper[4796]: I0930 16:26:01.305859 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-55476bd9c7-gvldb" Sep 30 16:26:03 crc kubenswrapper[4796]: I0930 16:26:03.310293 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" event={"ID":"4131038b-1bbe-410a-9c7b-58216c527106","Type":"ContainerStarted","Data":"6b5bab59d3a6556342f8e6498ae92538d57f68250c1b272875ed41116f5c5c1c"} Sep 30 16:26:03 crc kubenswrapper[4796]: I0930 16:26:03.312258 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" event={"ID":"a6b170fa-6665-405f-aaa3-042e9705ed1a","Type":"ContainerStarted","Data":"a10ef788fd31c3a305926d51848c0c1691b9dbdd95d5d03c8f91029f2c0773ac"} Sep 30 16:26:03 crc kubenswrapper[4796]: I0930 16:26:03.314714 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" event={"ID":"24427541-c81d-4542-b20e-bdcf3a4f0e4c","Type":"ContainerStarted","Data":"1a88fb42f4eef05499117d0aa0b6bf6334d4207fa60948c43c42bee3c52e6149"} Sep 30 16:26:03 crc kubenswrapper[4796]: I0930 16:26:03.322930 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" event={"ID":"e50b54b7-c7c3-4614-bd2e-3b90b68e129f","Type":"ContainerStarted","Data":"1d70e25362d17abd37c0a1ae77360662121d8663524523c2e0f10576877b3500"} Sep 30 16:26:03 crc kubenswrapper[4796]: I0930 16:26:03.338563 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" event={"ID":"ed0b77fc-1f2a-4a53-9983-ba64a6831569","Type":"ContainerStarted","Data":"f77b3a0e7d6c73e0db975e238b356c7fe260b11513df7e8407bc8606c35f962a"} Sep 30 16:26:03 crc kubenswrapper[4796]: I0930 16:26:03.348760 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" event={"ID":"8fba4c88-186e-4a06-8374-97e696008fcd","Type":"ContainerStarted","Data":"9a3c0c6f729e36bd9dcb28485042987d9f6c5d5f1295e94bd55d9653555a95c5"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.384464 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" event={"ID":"ae3399ed-5451-44ba-9872-73d709faab31","Type":"ContainerStarted","Data":"eb34469b226a96cf59e55f80a7832ddee7aac437583fcb3af2452785f5f801f9"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.384879 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" event={"ID":"ae3399ed-5451-44ba-9872-73d709faab31","Type":"ContainerStarted","Data":"2038d84769cef6ca4e17398b02beb6d849bb3c622a09076225ebd34ff26885cd"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.384898 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.387634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" event={"ID":"cf66eb4e-6595-40ea-b64a-b3d40e44dec9","Type":"ContainerStarted","Data":"86f5165b8ac556b8497475d42d41d6cb4d595a7c81c6fa567ee48f762ba59967"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.387675 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" event={"ID":"cf66eb4e-6595-40ea-b64a-b3d40e44dec9","Type":"ContainerStarted","Data":"86c1eee2bbaf4ffc5b5109d7d4fc3734d3083091478fbcc8bfc047e53c3bc239"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.387873 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.390482 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" event={"ID":"4131038b-1bbe-410a-9c7b-58216c527106","Type":"ContainerStarted","Data":"6f534cc4df42e1ce426a08fb63c4ae0006c01ff445eafe4d5d17c49c361badb2"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.390553 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.392178 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" event={"ID":"a6b170fa-6665-405f-aaa3-042e9705ed1a","Type":"ContainerStarted","Data":"aa6b50aadb4d49376f60fd64827cceef54d10efd82c971ca983bf3fc1d6ff4f2"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.392294 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.393588 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" event={"ID":"16086739-bea6-4c47-85d8-9a10497c9373","Type":"ContainerStarted","Data":"e0b2c622a562f476e9f80502163fba7816ebb546eb5a1136efd61e4d83ce03b4"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.393616 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" event={"ID":"16086739-bea6-4c47-85d8-9a10497c9373","Type":"ContainerStarted","Data":"737393da64c41cdde885c83b5d87e438815818f1aa24b65c317c6d965d6b2a7b"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.393964 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.397016 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" event={"ID":"24427541-c81d-4542-b20e-bdcf3a4f0e4c","Type":"ContainerStarted","Data":"e6b686b4d5f25066ceed0194f9d07ba341388f3038605a34e8364fd0f7d59d2f"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.397188 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.398665 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" event={"ID":"36b7f470-4ef5-431f-a756-04c40e1afc90","Type":"ContainerStarted","Data":"e6aeaca2aec114db6b7fc6d73fa6a99b9ce094d42a94768688d8b9a0d222b1d9"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.398691 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" event={"ID":"36b7f470-4ef5-431f-a756-04c40e1afc90","Type":"ContainerStarted","Data":"68c154b66c820760703262e5d0f5d7b3a5cd32a49e16f4e8032d8568f72baca8"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.398826 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.400525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" event={"ID":"e50b54b7-c7c3-4614-bd2e-3b90b68e129f","Type":"ContainerStarted","Data":"ef092695aa03e98e9d5ff8c11cdb955eaea8a95548adbb81b397c8abecd8d84c"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.400922 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.402258 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" event={"ID":"ed0b77fc-1f2a-4a53-9983-ba64a6831569","Type":"ContainerStarted","Data":"d67462281aaac0300c5c62577c6f7c07a461c1988e2c320c2b32d2aded401e19"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.402543 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.404177 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" podStartSLOduration=3.48350833 podStartE2EDuration="14.404155315s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.889615376 +0000 UTC m=+843.902893903" lastFinishedPulling="2025-09-30 16:26:02.810262361 +0000 UTC m=+854.823540888" observedRunningTime="2025-09-30 16:26:04.399774244 +0000 UTC m=+856.413052771" watchObservedRunningTime="2025-09-30 16:26:04.404155315 +0000 UTC m=+856.417433842" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.404649 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" event={"ID":"70f8a96e-36a5-4824-a041-72b7ccfbe064","Type":"ContainerStarted","Data":"7342d6f672106bffedc72457940a0e2d97f95f3555b751bf497b9a27e2ffb21f"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.404687 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" event={"ID":"70f8a96e-36a5-4824-a041-72b7ccfbe064","Type":"ContainerStarted","Data":"344bf7997d77ec66b9d5b5208cdcd1f84d796b4c45ffe0ee055454099b88e7ba"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.404820 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.408927 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" event={"ID":"8fba4c88-186e-4a06-8374-97e696008fcd","Type":"ContainerStarted","Data":"b70368bce8e20f6839a33caaa6196a8b08e139ec959961fc309e588d1df634b2"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.409056 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.411588 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" event={"ID":"d70c88cd-5ff3-49d9-b581-54860881ea39","Type":"ContainerStarted","Data":"2257320d20004a050f061482537e7fc2edfacde97af2a5a8799d4f25c8c77eb7"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.413524 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" event={"ID":"d1555f09-a29d-45bf-8bdf-5a3212d1ce0a","Type":"ContainerStarted","Data":"643a3c3ef7e4d605d5519d9e48c78054b06f310d6f412975365d790409e13492"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.413555 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" event={"ID":"d1555f09-a29d-45bf-8bdf-5a3212d1ce0a","Type":"ContainerStarted","Data":"510091fe6b8088d445a4b34314058f2d8ebaf4507c3bae010c247cdf7cbf65db"} Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.414113 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.415190 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" podStartSLOduration=4.515265074 podStartE2EDuration="15.415178066s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.865122621 +0000 UTC m=+843.878401148" lastFinishedPulling="2025-09-30 16:26:02.765035613 +0000 UTC m=+854.778314140" observedRunningTime="2025-09-30 16:26:04.414519237 +0000 UTC m=+856.427797764" watchObservedRunningTime="2025-09-30 16:26:04.415178066 +0000 UTC m=+856.428456593" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.450113 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" podStartSLOduration=4.43116601 podStartE2EDuration="15.450098035s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.79248831 +0000 UTC m=+843.805766837" lastFinishedPulling="2025-09-30 16:26:02.811420335 +0000 UTC m=+854.824698862" observedRunningTime="2025-09-30 16:26:04.44992863 +0000 UTC m=+856.463207157" watchObservedRunningTime="2025-09-30 16:26:04.450098035 +0000 UTC m=+856.463376562" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.497133 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" podStartSLOduration=4.593968659 podStartE2EDuration="15.497116127s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.861208014 +0000 UTC m=+843.874486541" lastFinishedPulling="2025-09-30 16:26:02.764355482 +0000 UTC m=+854.777634009" observedRunningTime="2025-09-30 16:26:04.479597191 +0000 UTC m=+856.492875718" watchObservedRunningTime="2025-09-30 16:26:04.497116127 +0000 UTC m=+856.510394654" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.497855 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" podStartSLOduration=4.592296808 podStartE2EDuration="15.497850619s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.860614026 +0000 UTC m=+843.873892553" lastFinishedPulling="2025-09-30 16:26:02.766167837 +0000 UTC m=+854.779446364" observedRunningTime="2025-09-30 16:26:04.494276121 +0000 UTC m=+856.507554638" watchObservedRunningTime="2025-09-30 16:26:04.497850619 +0000 UTC m=+856.511129146" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.513233 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" podStartSLOduration=4.169164513 podStartE2EDuration="15.51321603s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.415863122 +0000 UTC m=+843.429141649" lastFinishedPulling="2025-09-30 16:26:02.759914639 +0000 UTC m=+854.773193166" observedRunningTime="2025-09-30 16:26:04.506538309 +0000 UTC m=+856.519816836" watchObservedRunningTime="2025-09-30 16:26:04.51321603 +0000 UTC m=+856.526494557" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.528014 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" podStartSLOduration=4.652877127 podStartE2EDuration="15.527996914s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.890997388 +0000 UTC m=+843.904275915" lastFinishedPulling="2025-09-30 16:26:02.766117145 +0000 UTC m=+854.779395702" observedRunningTime="2025-09-30 16:26:04.525720675 +0000 UTC m=+856.538999202" watchObservedRunningTime="2025-09-30 16:26:04.527996914 +0000 UTC m=+856.541275441" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.547360 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" podStartSLOduration=4.530409579 podStartE2EDuration="15.547340354s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.793260354 +0000 UTC m=+843.806538881" lastFinishedPulling="2025-09-30 16:26:02.810191089 +0000 UTC m=+854.823469656" observedRunningTime="2025-09-30 16:26:04.54619396 +0000 UTC m=+856.559472507" watchObservedRunningTime="2025-09-30 16:26:04.547340354 +0000 UTC m=+856.560618881" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.570213 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" podStartSLOduration=4.524372819 podStartE2EDuration="15.570194911s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.769719197 +0000 UTC m=+843.782997724" lastFinishedPulling="2025-09-30 16:26:02.815541289 +0000 UTC m=+854.828819816" observedRunningTime="2025-09-30 16:26:04.565627864 +0000 UTC m=+856.578906391" watchObservedRunningTime="2025-09-30 16:26:04.570194911 +0000 UTC m=+856.583473438" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.589019 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-dwv6m" podStartSLOduration=3.819503789 podStartE2EDuration="14.589005025s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:52.139231491 +0000 UTC m=+844.152510018" lastFinishedPulling="2025-09-30 16:26:02.908732687 +0000 UTC m=+854.922011254" observedRunningTime="2025-09-30 16:26:04.585034156 +0000 UTC m=+856.598312683" watchObservedRunningTime="2025-09-30 16:26:04.589005025 +0000 UTC m=+856.602283552" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.604533 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" podStartSLOduration=4.583629537 podStartE2EDuration="15.604517191s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.744166139 +0000 UTC m=+843.757444656" lastFinishedPulling="2025-09-30 16:26:02.765053783 +0000 UTC m=+854.778332310" observedRunningTime="2025-09-30 16:26:04.601219302 +0000 UTC m=+856.614497829" watchObservedRunningTime="2025-09-30 16:26:04.604517191 +0000 UTC m=+856.617795718" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.648803 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" podStartSLOduration=4.756534388 podStartE2EDuration="15.64878084s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.912088071 +0000 UTC m=+843.925366608" lastFinishedPulling="2025-09-30 16:26:02.804334533 +0000 UTC m=+854.817613060" observedRunningTime="2025-09-30 16:26:04.643991556 +0000 UTC m=+856.657270083" watchObservedRunningTime="2025-09-30 16:26:04.64878084 +0000 UTC m=+856.662059367" Sep 30 16:26:04 crc kubenswrapper[4796]: I0930 16:26:04.650374 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" podStartSLOduration=4.245954739 podStartE2EDuration="15.650368178s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.42178137 +0000 UTC m=+843.435059897" lastFinishedPulling="2025-09-30 16:26:02.826194809 +0000 UTC m=+854.839473336" observedRunningTime="2025-09-30 16:26:04.621315856 +0000 UTC m=+856.634594383" watchObservedRunningTime="2025-09-30 16:26:04.650368178 +0000 UTC m=+856.663646705" Sep 30 16:26:09 crc kubenswrapper[4796]: I0930 16:26:09.460115 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" event={"ID":"52af31b2-ed79-46c4-840b-c2453d43a88c","Type":"ContainerStarted","Data":"deed1a20d2f1874c57ef05c481f63d328262b759f60f8295d83109f18af95c81"} Sep 30 16:26:09 crc kubenswrapper[4796]: I0930 16:26:09.461037 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:26:09 crc kubenswrapper[4796]: I0930 16:26:09.464028 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" event={"ID":"5cadf327-7ef7-439e-b692-5eef55c1b666","Type":"ContainerStarted","Data":"60ef937071bb68b83bed2aeba225618d08f36b7509d7c4a808fdd5283cc776d9"} Sep 30 16:26:09 crc kubenswrapper[4796]: I0930 16:26:09.464206 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:26:09 crc kubenswrapper[4796]: I0930 16:26:09.480463 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" podStartSLOduration=3.647577964 podStartE2EDuration="20.480444467s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.948550706 +0000 UTC m=+843.961829233" lastFinishedPulling="2025-09-30 16:26:08.781417209 +0000 UTC m=+860.794695736" observedRunningTime="2025-09-30 16:26:09.476467198 +0000 UTC m=+861.489745725" watchObservedRunningTime="2025-09-30 16:26:09.480444467 +0000 UTC m=+861.493722994" Sep 30 16:26:09 crc kubenswrapper[4796]: I0930 16:26:09.494999 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" podStartSLOduration=2.667891774 podStartE2EDuration="19.494962063s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.954960889 +0000 UTC m=+843.968239416" lastFinishedPulling="2025-09-30 16:26:08.782031188 +0000 UTC m=+860.795309705" observedRunningTime="2025-09-30 16:26:09.493988334 +0000 UTC m=+861.507266881" watchObservedRunningTime="2025-09-30 16:26:09.494962063 +0000 UTC m=+861.508240590" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.147551 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-nvqft" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.167038 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-4dsbl" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.180527 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-jxwnq" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.246529 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-j6vsj" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.287104 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fqxdk" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.380893 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-sb9j9" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.382347 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-dlpc6" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.407598 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-44kvk" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.542959 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-mttjp" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.570795 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-jj598" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.615288 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-q2knx" Sep 30 16:26:10 crc kubenswrapper[4796]: I0930 16:26:10.659093 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-whxjg" Sep 30 16:26:11 crc kubenswrapper[4796]: I0930 16:26:11.484563 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" event={"ID":"8ef65c41-fdac-4065-a568-2bd9f1176adc","Type":"ContainerStarted","Data":"e87feb8408573de381ea798f2cf2c92ca484bc9dc2032f717153e5d18521354a"} Sep 30 16:26:11 crc kubenswrapper[4796]: I0930 16:26:11.485819 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.493728 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" event={"ID":"2a580ec9-b0df-41bb-a272-cc4076cacdba","Type":"ContainerStarted","Data":"84754b3b3464a9f9654fa47d8cfeabfd7f8175399f04ca910324251899386539"} Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.494250 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.495923 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" event={"ID":"cb7865b9-6ef2-4cc1-92e5-0753a28e43c1","Type":"ContainerStarted","Data":"7f03da1a49a899f653df2e25f531e21249c9584d8c58c8b3fed766cbd470e177"} Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.496167 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.498089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" event={"ID":"530de9cb-0bdc-4773-a4ff-29b992d1942f","Type":"ContainerStarted","Data":"4066086d5f45661b959356b4f2ee80c605177e899fcf31a55db9c1f447b04606"} Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.498283 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.501108 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" event={"ID":"5a8c84fa-799a-4f47-81c1-7445702e0f23","Type":"ContainerStarted","Data":"93f2f944e9f6b16c32a39ed185d2f31a61bf95c396394858579a0ffe2a60db20"} Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.501753 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.503552 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" event={"ID":"df653968-5944-4e8d-9e82-7e19a539d997","Type":"ContainerStarted","Data":"46cf8628bf6f703846b8071ed2c3ffab314734ec1ebd3a373fb92b507f788573"} Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.503973 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.505662 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" event={"ID":"40fc5188-af1b-454f-b7f4-6c1ae1c4cf71","Type":"ContainerStarted","Data":"e359142f3ee8bfdb83983e44dbb8f6c3588873f3bef7f3bda392fd964816c47e"} Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.520676 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" podStartSLOduration=6.911517471 podStartE2EDuration="23.520659287s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:52.197209512 +0000 UTC m=+844.210488039" lastFinishedPulling="2025-09-30 16:26:08.806351328 +0000 UTC m=+860.819629855" observedRunningTime="2025-09-30 16:26:11.506333293 +0000 UTC m=+863.519611820" watchObservedRunningTime="2025-09-30 16:26:12.520659287 +0000 UTC m=+864.533937814" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.522056 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" podStartSLOduration=3.858140711 podStartE2EDuration="22.522050399s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:52.194968125 +0000 UTC m=+844.208246652" lastFinishedPulling="2025-09-30 16:26:10.858877813 +0000 UTC m=+862.872156340" observedRunningTime="2025-09-30 16:26:12.517603486 +0000 UTC m=+864.530882013" watchObservedRunningTime="2025-09-30 16:26:12.522050399 +0000 UTC m=+864.535328926" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.541326 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" podStartSLOduration=3.460099085 podStartE2EDuration="23.541302927s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.928566066 +0000 UTC m=+843.941844603" lastFinishedPulling="2025-09-30 16:26:12.009769918 +0000 UTC m=+864.023048445" observedRunningTime="2025-09-30 16:26:12.536375209 +0000 UTC m=+864.549653736" watchObservedRunningTime="2025-09-30 16:26:12.541302927 +0000 UTC m=+864.554581444" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.561926 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" podStartSLOduration=3.739164885 podStartE2EDuration="23.561910466s" podCreationTimestamp="2025-09-30 16:25:49 +0000 UTC" firstStartedPulling="2025-09-30 16:25:52.196430919 +0000 UTC m=+844.209709446" lastFinishedPulling="2025-09-30 16:26:12.0191765 +0000 UTC m=+864.032455027" observedRunningTime="2025-09-30 16:26:12.559020309 +0000 UTC m=+864.572298836" watchObservedRunningTime="2025-09-30 16:26:12.561910466 +0000 UTC m=+864.575188993" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.579620 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" podStartSLOduration=2.509808447 podStartE2EDuration="22.579602147s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:51.948863485 +0000 UTC m=+843.962142012" lastFinishedPulling="2025-09-30 16:26:12.018657185 +0000 UTC m=+864.031935712" observedRunningTime="2025-09-30 16:26:12.577793673 +0000 UTC m=+864.591072200" watchObservedRunningTime="2025-09-30 16:26:12.579602147 +0000 UTC m=+864.592880674" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.598352 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" podStartSLOduration=2.760120042 podStartE2EDuration="22.598333619s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:52.16918655 +0000 UTC m=+844.182465077" lastFinishedPulling="2025-09-30 16:26:12.007400117 +0000 UTC m=+864.020678654" observedRunningTime="2025-09-30 16:26:12.593827844 +0000 UTC m=+864.607106371" watchObservedRunningTime="2025-09-30 16:26:12.598333619 +0000 UTC m=+864.611612146" Sep 30 16:26:12 crc kubenswrapper[4796]: I0930 16:26:12.618555 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" podStartSLOduration=2.748146814 podStartE2EDuration="22.618525776s" podCreationTimestamp="2025-09-30 16:25:50 +0000 UTC" firstStartedPulling="2025-09-30 16:25:52.165197651 +0000 UTC m=+844.178476178" lastFinishedPulling="2025-09-30 16:26:12.035576603 +0000 UTC m=+864.048855140" observedRunningTime="2025-09-30 16:26:12.615146574 +0000 UTC m=+864.628425101" watchObservedRunningTime="2025-09-30 16:26:12.618525776 +0000 UTC m=+864.631804303" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.331130 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-j8d9f" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.393340 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-nmqln" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.634771 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-jqbdb" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.679718 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-f45lz" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.833531 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-97m97" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.901598 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6db74cfd6d-l8qfg" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.926360 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.930185 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-f66b554c6-dntbx" Sep 30 16:26:20 crc kubenswrapper[4796]: I0930 16:26:20.954740 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-j4ntp" Sep 30 16:26:21 crc kubenswrapper[4796]: I0930 16:26:21.206101 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p2jwc" Sep 30 16:26:31 crc kubenswrapper[4796]: I0930 16:26:31.097502 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:26:31 crc kubenswrapper[4796]: I0930 16:26:31.098309 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.725498 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9v9"] Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.728091 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.730953 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.731290 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.731740 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-2rf45" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.733843 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.768222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5205be66-cc23-4858-969e-b54efa4e6695-config\") pod \"dnsmasq-dns-675f4bcbfc-xh9v9\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.768324 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p96f\" (UniqueName: \"kubernetes.io/projected/5205be66-cc23-4858-969e-b54efa4e6695-kube-api-access-5p96f\") pod \"dnsmasq-dns-675f4bcbfc-xh9v9\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.813597 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9v9"] Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.816432 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zrnlt"] Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.818453 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.822928 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zrnlt"] Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.824237 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.869864 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5205be66-cc23-4858-969e-b54efa4e6695-config\") pod \"dnsmasq-dns-675f4bcbfc-xh9v9\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.870463 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-config\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.870684 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fscg\" (UniqueName: \"kubernetes.io/projected/43fd48e6-cf80-474a-a582-ea693e5e1e16-kube-api-access-6fscg\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.870719 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p96f\" (UniqueName: \"kubernetes.io/projected/5205be66-cc23-4858-969e-b54efa4e6695-kube-api-access-5p96f\") pod \"dnsmasq-dns-675f4bcbfc-xh9v9\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.870749 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.870806 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5205be66-cc23-4858-969e-b54efa4e6695-config\") pod \"dnsmasq-dns-675f4bcbfc-xh9v9\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.909233 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p96f\" (UniqueName: \"kubernetes.io/projected/5205be66-cc23-4858-969e-b54efa4e6695-kube-api-access-5p96f\") pod \"dnsmasq-dns-675f4bcbfc-xh9v9\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.971756 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.971874 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-config\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.971917 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fscg\" (UniqueName: \"kubernetes.io/projected/43fd48e6-cf80-474a-a582-ea693e5e1e16-kube-api-access-6fscg\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.972806 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.972871 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-config\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:38 crc kubenswrapper[4796]: I0930 16:26:38.989266 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fscg\" (UniqueName: \"kubernetes.io/projected/43fd48e6-cf80-474a-a582-ea693e5e1e16-kube-api-access-6fscg\") pod \"dnsmasq-dns-78dd6ddcc-zrnlt\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:39 crc kubenswrapper[4796]: I0930 16:26:39.111126 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:39 crc kubenswrapper[4796]: I0930 16:26:39.147551 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:39 crc kubenswrapper[4796]: I0930 16:26:39.607825 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9v9"] Sep 30 16:26:39 crc kubenswrapper[4796]: I0930 16:26:39.645382 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zrnlt"] Sep 30 16:26:39 crc kubenswrapper[4796]: W0930 16:26:39.652791 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43fd48e6_cf80_474a_a582_ea693e5e1e16.slice/crio-d6f15c504bbd4ef17e48f72e853b9a29c4c9a86d7709ad5293c797202dcede6e WatchSource:0}: Error finding container d6f15c504bbd4ef17e48f72e853b9a29c4c9a86d7709ad5293c797202dcede6e: Status 404 returned error can't find the container with id d6f15c504bbd4ef17e48f72e853b9a29c4c9a86d7709ad5293c797202dcede6e Sep 30 16:26:39 crc kubenswrapper[4796]: I0930 16:26:39.766127 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" event={"ID":"43fd48e6-cf80-474a-a582-ea693e5e1e16","Type":"ContainerStarted","Data":"d6f15c504bbd4ef17e48f72e853b9a29c4c9a86d7709ad5293c797202dcede6e"} Sep 30 16:26:39 crc kubenswrapper[4796]: I0930 16:26:39.768060 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" event={"ID":"5205be66-cc23-4858-969e-b54efa4e6695","Type":"ContainerStarted","Data":"f0511a10c9fe3198fc4ef12b3fd5db9205aff110428e2e8a933305734b29d417"} Sep 30 16:26:41 crc kubenswrapper[4796]: I0930 16:26:41.836143 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9v9"] Sep 30 16:26:41 crc kubenswrapper[4796]: I0930 16:26:41.905322 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-hpsg7"] Sep 30 16:26:41 crc kubenswrapper[4796]: I0930 16:26:41.907071 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:41 crc kubenswrapper[4796]: I0930 16:26:41.919317 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-hpsg7"] Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.121536 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-config\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.121664 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-dns-svc\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.121728 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfzbk\" (UniqueName: \"kubernetes.io/projected/11ffb893-0b90-4b17-9a48-a67f16eca6ac-kube-api-access-nfzbk\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.143185 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zrnlt"] Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.171334 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bsdtj"] Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.172654 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.185502 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bsdtj"] Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.225639 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-config\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.226163 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-dns-svc\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.226855 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfzbk\" (UniqueName: \"kubernetes.io/projected/11ffb893-0b90-4b17-9a48-a67f16eca6ac-kube-api-access-nfzbk\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.227224 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-dns-svc\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.227283 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-config\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.253357 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfzbk\" (UniqueName: \"kubernetes.io/projected/11ffb893-0b90-4b17-9a48-a67f16eca6ac-kube-api-access-nfzbk\") pod \"dnsmasq-dns-666b6646f7-hpsg7\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.329880 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvkqb\" (UniqueName: \"kubernetes.io/projected/e6909bbc-1c68-43ff-afd5-6fbb1078297c-kube-api-access-rvkqb\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.329935 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-config\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.330015 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.430908 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvkqb\" (UniqueName: \"kubernetes.io/projected/e6909bbc-1c68-43ff-afd5-6fbb1078297c-kube-api-access-rvkqb\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.430955 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-config\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.431042 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.432684 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.432697 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-config\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.451199 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvkqb\" (UniqueName: \"kubernetes.io/projected/e6909bbc-1c68-43ff-afd5-6fbb1078297c-kube-api-access-rvkqb\") pod \"dnsmasq-dns-57d769cc4f-bsdtj\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.511371 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:26:42 crc kubenswrapper[4796]: I0930 16:26:42.542405 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.009449 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.011366 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.019082 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.019375 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.019519 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.021805 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.023882 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.025302 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.025676 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.025929 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-khhtv" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141507 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141591 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-config-data\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141638 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141659 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141681 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141701 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fw79\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-kube-api-access-7fw79\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141724 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141778 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141799 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.141816 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243648 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243704 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243729 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fw79\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-kube-api-access-7fw79\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243756 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243777 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243837 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243854 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243879 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.243910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-config-data\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.244016 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.245016 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.245300 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.245761 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.246279 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.246790 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-config-data\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.248066 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.249937 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.250310 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.250827 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.259251 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.261654 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fw79\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-kube-api-access-7fw79\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.268639 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.298348 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.299590 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302509 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gwtqv" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302669 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302529 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302600 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302612 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302641 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.302660 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.315637 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.368275 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447358 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnj2s\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-kube-api-access-mnj2s\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447433 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447457 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447482 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447514 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447557 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447587 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447644 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.447667 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549046 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549126 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549155 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549203 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549228 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549252 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549277 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549297 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549335 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549550 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549582 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnj2s\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-kube-api-access-mnj2s\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.549895 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.550431 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.552967 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.553634 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.556782 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.558538 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.558912 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.559803 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.561258 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.570916 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnj2s\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-kube-api-access-mnj2s\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.579420 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:43 crc kubenswrapper[4796]: I0930 16:26:43.645259 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.988971 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.991612 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.995014 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.995331 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.995348 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.995946 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.996269 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 30 16:26:45 crc kubenswrapper[4796]: I0930 16:26:45.996891 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6llds" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.002298 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095105 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095468 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-config-data-default\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095499 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095550 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095573 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095756 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.095819 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-secrets\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.096062 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlhjr\" (UniqueName: \"kubernetes.io/projected/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-kube-api-access-qlhjr\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.096190 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-kolla-config\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.114669 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.116308 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.125251 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wwphj" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.126305 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.126373 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.126552 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.142818 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.198662 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.198748 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2bb8f271-8840-42e7-984c-31cf8fc8bea0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.198775 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.198915 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.198953 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199015 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199071 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-config-data-default\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199101 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199161 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199189 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n8qq\" (UniqueName: \"kubernetes.io/projected/2bb8f271-8840-42e7-984c-31cf8fc8bea0-kube-api-access-2n8qq\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199242 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199267 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199317 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199353 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199378 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-secrets\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199443 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199477 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlhjr\" (UniqueName: \"kubernetes.io/projected/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-kube-api-access-qlhjr\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.199505 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-kolla-config\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.200082 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.200140 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.200477 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-kolla-config\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.200666 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.204785 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-config-data-default\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.209445 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.211568 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-secrets\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.226690 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlhjr\" (UniqueName: \"kubernetes.io/projected/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-kube-api-access-qlhjr\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.227255 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6162abb2-dddc-4cb6-bd50-0e71ae9534bd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.239744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6162abb2-dddc-4cb6-bd50-0e71ae9534bd\") " pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300741 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300798 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2bb8f271-8840-42e7-984c-31cf8fc8bea0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300848 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300885 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300940 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.300964 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n8qq\" (UniqueName: \"kubernetes.io/projected/2bb8f271-8840-42e7-984c-31cf8fc8bea0-kube-api-access-2n8qq\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.301022 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.301074 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.301531 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.302303 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2bb8f271-8840-42e7-984c-31cf8fc8bea0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.302336 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.302642 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.303263 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bb8f271-8840-42e7-984c-31cf8fc8bea0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.314584 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.317726 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.318539 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb8f271-8840-42e7-984c-31cf8fc8bea0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.323200 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.327733 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n8qq\" (UniqueName: \"kubernetes.io/projected/2bb8f271-8840-42e7-984c-31cf8fc8bea0-kube-api-access-2n8qq\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.342002 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2bb8f271-8840-42e7-984c-31cf8fc8bea0\") " pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.394003 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.394958 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.398760 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.398968 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-g9wsg" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.399087 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.404702 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.452045 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.504128 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djdw5\" (UniqueName: \"kubernetes.io/projected/94a4df46-32b9-4a28-848e-4c90a2660b82-kube-api-access-djdw5\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.504179 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a4df46-32b9-4a28-848e-4c90a2660b82-combined-ca-bundle\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.504303 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94a4df46-32b9-4a28-848e-4c90a2660b82-kolla-config\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.504340 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a4df46-32b9-4a28-848e-4c90a2660b82-memcached-tls-certs\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.504363 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94a4df46-32b9-4a28-848e-4c90a2660b82-config-data\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.605795 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94a4df46-32b9-4a28-848e-4c90a2660b82-kolla-config\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.605864 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a4df46-32b9-4a28-848e-4c90a2660b82-memcached-tls-certs\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.605901 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94a4df46-32b9-4a28-848e-4c90a2660b82-config-data\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.605933 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djdw5\" (UniqueName: \"kubernetes.io/projected/94a4df46-32b9-4a28-848e-4c90a2660b82-kube-api-access-djdw5\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.605970 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a4df46-32b9-4a28-848e-4c90a2660b82-combined-ca-bundle\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.606847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94a4df46-32b9-4a28-848e-4c90a2660b82-config-data\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.607148 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94a4df46-32b9-4a28-848e-4c90a2660b82-kolla-config\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.609539 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a4df46-32b9-4a28-848e-4c90a2660b82-combined-ca-bundle\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.609572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a4df46-32b9-4a28-848e-4c90a2660b82-memcached-tls-certs\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.631890 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djdw5\" (UniqueName: \"kubernetes.io/projected/94a4df46-32b9-4a28-848e-4c90a2660b82-kube-api-access-djdw5\") pod \"memcached-0\" (UID: \"94a4df46-32b9-4a28-848e-4c90a2660b82\") " pod="openstack/memcached-0" Sep 30 16:26:46 crc kubenswrapper[4796]: I0930 16:26:46.712745 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.085065 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.086212 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.088637 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qnlx8" Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.099431 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.131097 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25qtl\" (UniqueName: \"kubernetes.io/projected/f340e3c8-c819-488b-808a-f610a2b60ef7-kube-api-access-25qtl\") pod \"kube-state-metrics-0\" (UID: \"f340e3c8-c819-488b-808a-f610a2b60ef7\") " pod="openstack/kube-state-metrics-0" Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.232973 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25qtl\" (UniqueName: \"kubernetes.io/projected/f340e3c8-c819-488b-808a-f610a2b60ef7-kube-api-access-25qtl\") pod \"kube-state-metrics-0\" (UID: \"f340e3c8-c819-488b-808a-f610a2b60ef7\") " pod="openstack/kube-state-metrics-0" Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.273749 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25qtl\" (UniqueName: \"kubernetes.io/projected/f340e3c8-c819-488b-808a-f610a2b60ef7-kube-api-access-25qtl\") pod \"kube-state-metrics-0\" (UID: \"f340e3c8-c819-488b-808a-f610a2b60ef7\") " pod="openstack/kube-state-metrics-0" Sep 30 16:26:48 crc kubenswrapper[4796]: I0930 16:26:48.418557 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.261566 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vfdx6"] Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.266265 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.269497 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.269803 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-mlhwn" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.269940 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.275380 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vfdx6"] Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.311839 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-lg6m7"] Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.313508 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.330113 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lg6m7"] Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.383788 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh4p2\" (UniqueName: \"kubernetes.io/projected/713f98aa-c3ec-4627-bc90-a1a577ca1c27-kube-api-access-hh4p2\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.383838 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-run\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.383882 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/713f98aa-c3ec-4627-bc90-a1a577ca1c27-ovn-controller-tls-certs\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.383911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-run-ovn\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.384739 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713f98aa-c3ec-4627-bc90-a1a577ca1c27-combined-ca-bundle\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.385138 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-log-ovn\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.385203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/713f98aa-c3ec-4627-bc90-a1a577ca1c27-scripts\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.486671 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-run\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.486717 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-log-ovn\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.486783 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/713f98aa-c3ec-4627-bc90-a1a577ca1c27-scripts\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.486830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-etc-ovs\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.486872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh4p2\" (UniqueName: \"kubernetes.io/projected/713f98aa-c3ec-4627-bc90-a1a577ca1c27-kube-api-access-hh4p2\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.486897 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-run\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487072 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/403ec30e-d599-4940-8eaf-6cccb3200ed9-scripts\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487107 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/713f98aa-c3ec-4627-bc90-a1a577ca1c27-ovn-controller-tls-certs\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487144 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-run-ovn\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487177 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-log\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487199 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-lib\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487236 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpttp\" (UniqueName: \"kubernetes.io/projected/403ec30e-d599-4940-8eaf-6cccb3200ed9-kube-api-access-mpttp\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487279 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713f98aa-c3ec-4627-bc90-a1a577ca1c27-combined-ca-bundle\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487649 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-run\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487696 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-run-ovn\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.487737 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/713f98aa-c3ec-4627-bc90-a1a577ca1c27-var-log-ovn\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.489117 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/713f98aa-c3ec-4627-bc90-a1a577ca1c27-scripts\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.494016 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713f98aa-c3ec-4627-bc90-a1a577ca1c27-combined-ca-bundle\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.495928 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/713f98aa-c3ec-4627-bc90-a1a577ca1c27-ovn-controller-tls-certs\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.509386 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh4p2\" (UniqueName: \"kubernetes.io/projected/713f98aa-c3ec-4627-bc90-a1a577ca1c27-kube-api-access-hh4p2\") pod \"ovn-controller-vfdx6\" (UID: \"713f98aa-c3ec-4627-bc90-a1a577ca1c27\") " pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.587940 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vfdx6" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.588676 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-etc-ovs\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.588831 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/403ec30e-d599-4940-8eaf-6cccb3200ed9-scripts\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.588919 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-log\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.588957 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-lib\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.589120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpttp\" (UniqueName: \"kubernetes.io/projected/403ec30e-d599-4940-8eaf-6cccb3200ed9-kube-api-access-mpttp\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.589228 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-run\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.589397 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-log\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.589447 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-run\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.589540 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-etc-ovs\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.589787 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/403ec30e-d599-4940-8eaf-6cccb3200ed9-var-lib\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.591409 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/403ec30e-d599-4940-8eaf-6cccb3200ed9-scripts\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.624878 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpttp\" (UniqueName: \"kubernetes.io/projected/403ec30e-d599-4940-8eaf-6cccb3200ed9-kube-api-access-mpttp\") pod \"ovn-controller-ovs-lg6m7\" (UID: \"403ec30e-d599-4940-8eaf-6cccb3200ed9\") " pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:51 crc kubenswrapper[4796]: I0930 16:26:51.642825 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.575912 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.578422 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.580586 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.580913 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.581254 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-qm7g4" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.581333 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.586759 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.588965 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.739838 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.739893 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.739915 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.740060 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.740090 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-config\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.740112 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lh8z\" (UniqueName: \"kubernetes.io/projected/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-kube-api-access-4lh8z\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.740143 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.740170 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.844747 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845214 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845271 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-config\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845343 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lh8z\" (UniqueName: \"kubernetes.io/projected/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-kube-api-access-4lh8z\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845433 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845511 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845694 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.845748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.846759 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-config\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.846939 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.847497 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.852918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.853396 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.854267 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: E0930 16:26:53.867715 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 30 16:26:53 crc kubenswrapper[4796]: E0930 16:26:53.867916 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5p96f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-xh9v9_openstack(5205be66-cc23-4858-969e-b54efa4e6695): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 16:26:53 crc kubenswrapper[4796]: E0930 16:26:53.870245 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" podUID="5205be66-cc23-4858-969e-b54efa4e6695" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.872401 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lh8z\" (UniqueName: \"kubernetes.io/projected/85b72cb4-9a34-4d07-8b11-dc60f4d540c9-kube-api-access-4lh8z\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.873330 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85b72cb4-9a34-4d07-8b11-dc60f4d540c9\") " pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:53 crc kubenswrapper[4796]: E0930 16:26:53.899275 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 30 16:26:53 crc kubenswrapper[4796]: E0930 16:26:53.899448 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6fscg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zrnlt_openstack(43fd48e6-cf80-474a-a582-ea693e5e1e16): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 16:26:53 crc kubenswrapper[4796]: E0930 16:26:53.901079 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" podUID="43fd48e6-cf80-474a-a582-ea693e5e1e16" Sep 30 16:26:53 crc kubenswrapper[4796]: I0930 16:26:53.906739 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.433078 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.438788 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.555703 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5205be66-cc23-4858-969e-b54efa4e6695-config\") pod \"5205be66-cc23-4858-969e-b54efa4e6695\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.555888 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-config\") pod \"43fd48e6-cf80-474a-a582-ea693e5e1e16\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.556671 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-config" (OuterVolumeSpecName: "config") pod "43fd48e6-cf80-474a-a582-ea693e5e1e16" (UID: "43fd48e6-cf80-474a-a582-ea693e5e1e16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.556765 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fscg\" (UniqueName: \"kubernetes.io/projected/43fd48e6-cf80-474a-a582-ea693e5e1e16-kube-api-access-6fscg\") pod \"43fd48e6-cf80-474a-a582-ea693e5e1e16\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.557860 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-dns-svc\") pod \"43fd48e6-cf80-474a-a582-ea693e5e1e16\" (UID: \"43fd48e6-cf80-474a-a582-ea693e5e1e16\") " Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.557912 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p96f\" (UniqueName: \"kubernetes.io/projected/5205be66-cc23-4858-969e-b54efa4e6695-kube-api-access-5p96f\") pod \"5205be66-cc23-4858-969e-b54efa4e6695\" (UID: \"5205be66-cc23-4858-969e-b54efa4e6695\") " Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.558540 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43fd48e6-cf80-474a-a582-ea693e5e1e16" (UID: "43fd48e6-cf80-474a-a582-ea693e5e1e16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.558746 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.558784 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43fd48e6-cf80-474a-a582-ea693e5e1e16-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.559326 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5205be66-cc23-4858-969e-b54efa4e6695-config" (OuterVolumeSpecName: "config") pod "5205be66-cc23-4858-969e-b54efa4e6695" (UID: "5205be66-cc23-4858-969e-b54efa4e6695"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.575020 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43fd48e6-cf80-474a-a582-ea693e5e1e16-kube-api-access-6fscg" (OuterVolumeSpecName: "kube-api-access-6fscg") pod "43fd48e6-cf80-474a-a582-ea693e5e1e16" (UID: "43fd48e6-cf80-474a-a582-ea693e5e1e16"). InnerVolumeSpecName "kube-api-access-6fscg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.577173 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5205be66-cc23-4858-969e-b54efa4e6695-kube-api-access-5p96f" (OuterVolumeSpecName: "kube-api-access-5p96f") pod "5205be66-cc23-4858-969e-b54efa4e6695" (UID: "5205be66-cc23-4858-969e-b54efa4e6695"). InnerVolumeSpecName "kube-api-access-5p96f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.660681 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p96f\" (UniqueName: \"kubernetes.io/projected/5205be66-cc23-4858-969e-b54efa4e6695-kube-api-access-5p96f\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.660922 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5205be66-cc23-4858-969e-b54efa4e6695-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.660934 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fscg\" (UniqueName: \"kubernetes.io/projected/43fd48e6-cf80-474a-a582-ea693e5e1e16-kube-api-access-6fscg\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:54 crc kubenswrapper[4796]: W0930 16:26:54.765133 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod903f4e99_1ea4_46b3_8b5e_2ba439c5ba73.slice/crio-80aa571d96bf60b76d3701aa197d8c928f71f3e64058198ffd4f98c24b84ff2e WatchSource:0}: Error finding container 80aa571d96bf60b76d3701aa197d8c928f71f3e64058198ffd4f98c24b84ff2e: Status 404 returned error can't find the container with id 80aa571d96bf60b76d3701aa197d8c928f71f3e64058198ffd4f98c24b84ff2e Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.784131 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.799070 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: W0930 16:26:54.807926 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11ffb893_0b90_4b17_9a48_a67f16eca6ac.slice/crio-93d16dade0d2dd19fa9f7bbe1fcb7dd7ec137b77fb39bf1d62a978a7a4ce68bf WatchSource:0}: Error finding container 93d16dade0d2dd19fa9f7bbe1fcb7dd7ec137b77fb39bf1d62a978a7a4ce68bf: Status 404 returned error can't find the container with id 93d16dade0d2dd19fa9f7bbe1fcb7dd7ec137b77fb39bf1d62a978a7a4ce68bf Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.815362 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bsdtj"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.823561 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.830135 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.865246 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-hpsg7"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.879749 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.895949 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.900235 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vfdx6"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.931359 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73","Type":"ContainerStarted","Data":"80aa571d96bf60b76d3701aa197d8c928f71f3e64058198ffd4f98c24b84ff2e"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.932796 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vfdx6" event={"ID":"713f98aa-c3ec-4627-bc90-a1a577ca1c27","Type":"ContainerStarted","Data":"51fba3ac3672704da428bc98013aa06d2fb0d16954ecebd4833ebc114bdaf605"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.934190 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" event={"ID":"43fd48e6-cf80-474a-a582-ea693e5e1e16","Type":"ContainerDied","Data":"d6f15c504bbd4ef17e48f72e853b9a29c4c9a86d7709ad5293c797202dcede6e"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.934227 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zrnlt" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.935691 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" event={"ID":"5205be66-cc23-4858-969e-b54efa4e6695","Type":"ContainerDied","Data":"f0511a10c9fe3198fc4ef12b3fd5db9205aff110428e2e8a933305734b29d417"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.935922 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9v9" Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.936672 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2bb8f271-8840-42e7-984c-31cf8fc8bea0","Type":"ContainerStarted","Data":"7f51f3d19d94ecb040e50771a86879761487ec4592c48beeb075a98c6c5f5de2"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.937482 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" event={"ID":"e6909bbc-1c68-43ff-afd5-6fbb1078297c","Type":"ContainerStarted","Data":"ac041d45eecc4be46f0a04fe57f1761108c4292bbd35e09ddb6c93961dd64af7"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.940530 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6162abb2-dddc-4cb6-bd50-0e71ae9534bd","Type":"ContainerStarted","Data":"cf3d02bae8ba6548bc9173537a6e74f30a8bb1fc1373dcfa60a582f4a604ad91"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.941550 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f340e3c8-c819-488b-808a-f610a2b60ef7","Type":"ContainerStarted","Data":"149a83074c797ec61716f77f3a3ea7b5e4c1b73dae29f799ca2c14fb83865379"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.943478 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"94a4df46-32b9-4a28-848e-4c90a2660b82","Type":"ContainerStarted","Data":"06877acd3fb496735c09eb23a1a5888fa9ec26505300995bcc7253d75ec1805e"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.944237 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee37a44-c6b6-4d75-ad41-9ff1402925f1","Type":"ContainerStarted","Data":"2c7e94077a9e870b219f9a15767d5dfed0fc2a3c9697a87652be5d8c549b6179"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.944901 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" event={"ID":"11ffb893-0b90-4b17-9a48-a67f16eca6ac","Type":"ContainerStarted","Data":"93d16dade0d2dd19fa9f7bbe1fcb7dd7ec137b77fb39bf1d62a978a7a4ce68bf"} Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.975182 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9v9"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.981371 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 16:26:54 crc kubenswrapper[4796]: I0930 16:26:54.990784 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9v9"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.002512 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zrnlt"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.007303 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zrnlt"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.069400 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lg6m7"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.206494 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-pmjrt"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.207540 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.209888 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.216161 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-pmjrt"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.376421 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-ovn-rundir\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.376731 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-ovs-rundir\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.376752 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r564x\" (UniqueName: \"kubernetes.io/projected/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-kube-api-access-r564x\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.376791 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-config\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.376819 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.376842 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-combined-ca-bundle\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.478650 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-ovn-rundir\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.478731 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-ovs-rundir\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.478750 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r564x\" (UniqueName: \"kubernetes.io/projected/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-kube-api-access-r564x\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.478786 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-config\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.478815 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.478839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-combined-ca-bundle\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.479023 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-ovs-rundir\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.479095 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-ovn-rundir\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.480795 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-config\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.483823 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.484256 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-combined-ca-bundle\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.501372 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r564x\" (UniqueName: \"kubernetes.io/projected/f5bcbf15-75a7-4b1f-bb3d-4d58f59de746-kube-api-access-r564x\") pod \"ovn-controller-metrics-pmjrt\" (UID: \"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746\") " pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.542379 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-pmjrt" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.662896 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-hpsg7"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.701357 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kq8hq"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.702700 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.709258 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.711587 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kq8hq"] Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.888209 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.888257 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-config\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.888311 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncgzk\" (UniqueName: \"kubernetes.io/projected/11b92058-788c-40df-a406-afebc51358ca-kube-api-access-ncgzk\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.888339 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.954359 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lg6m7" event={"ID":"403ec30e-d599-4940-8eaf-6cccb3200ed9","Type":"ContainerStarted","Data":"70db672bb2d09798f59245d17377a24120fe95597b0b197ee088a2cfca8ca8cf"} Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.955987 4796 generic.go:334] "Generic (PLEG): container finished" podID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerID="f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb" exitCode=0 Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.956201 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" event={"ID":"e6909bbc-1c68-43ff-afd5-6fbb1078297c","Type":"ContainerDied","Data":"f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb"} Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.959413 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"85b72cb4-9a34-4d07-8b11-dc60f4d540c9","Type":"ContainerStarted","Data":"dbbaf3107c3c621dac3d0d4d978a401baa0a98e0dc95d065b89b87a8db27639d"} Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.968357 4796 generic.go:334] "Generic (PLEG): container finished" podID="11ffb893-0b90-4b17-9a48-a67f16eca6ac" containerID="f1d58ca672e086cc6776ef863690765aa30e07a35ac3fe61ce4503a0900d9e0b" exitCode=0 Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.968399 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" event={"ID":"11ffb893-0b90-4b17-9a48-a67f16eca6ac","Type":"ContainerDied","Data":"f1d58ca672e086cc6776ef863690765aa30e07a35ac3fe61ce4503a0900d9e0b"} Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.989496 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncgzk\" (UniqueName: \"kubernetes.io/projected/11b92058-788c-40df-a406-afebc51358ca-kube-api-access-ncgzk\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.989561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.989659 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.989702 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-config\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.991489 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.991996 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-config\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:55 crc kubenswrapper[4796]: I0930 16:26:55.992508 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.008455 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncgzk\" (UniqueName: \"kubernetes.io/projected/11b92058-788c-40df-a406-afebc51358ca-kube-api-access-ncgzk\") pod \"dnsmasq-dns-7fd796d7df-kq8hq\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.103317 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.359744 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.362870 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.364787 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.365341 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-t7glz" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.365348 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.365503 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.369238 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.439102 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.497765 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.497827 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-config\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.497850 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.497878 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9jdf\" (UniqueName: \"kubernetes.io/projected/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-kube-api-access-s9jdf\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.497904 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.497967 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.498044 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.498079 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.598714 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-config\") pod \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.598750 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-dns-svc\") pod \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.598792 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfzbk\" (UniqueName: \"kubernetes.io/projected/11ffb893-0b90-4b17-9a48-a67f16eca6ac-kube-api-access-nfzbk\") pod \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\" (UID: \"11ffb893-0b90-4b17-9a48-a67f16eca6ac\") " Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599019 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-config\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599045 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599069 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9jdf\" (UniqueName: \"kubernetes.io/projected/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-kube-api-access-s9jdf\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599100 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599165 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599214 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599244 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599265 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.599727 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.600425 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.601615 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-config\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.602963 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.613350 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.613543 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.615569 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ffb893-0b90-4b17-9a48-a67f16eca6ac-kube-api-access-nfzbk" (OuterVolumeSpecName: "kube-api-access-nfzbk") pod "11ffb893-0b90-4b17-9a48-a67f16eca6ac" (UID: "11ffb893-0b90-4b17-9a48-a67f16eca6ac"). InnerVolumeSpecName "kube-api-access-nfzbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.617656 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.621279 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9jdf\" (UniqueName: \"kubernetes.io/projected/a121b29f-66c2-4b9f-99b2-9a6c4a46473f-kube-api-access-s9jdf\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.633070 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11ffb893-0b90-4b17-9a48-a67f16eca6ac" (UID: "11ffb893-0b90-4b17-9a48-a67f16eca6ac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.637469 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a121b29f-66c2-4b9f-99b2-9a6c4a46473f\") " pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.639966 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-config" (OuterVolumeSpecName: "config") pod "11ffb893-0b90-4b17-9a48-a67f16eca6ac" (UID: "11ffb893-0b90-4b17-9a48-a67f16eca6ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.700785 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.700818 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ffb893-0b90-4b17-9a48-a67f16eca6ac-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.700828 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfzbk\" (UniqueName: \"kubernetes.io/projected/11ffb893-0b90-4b17-9a48-a67f16eca6ac-kube-api-access-nfzbk\") on node \"crc\" DevicePath \"\"" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.736581 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.743965 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43fd48e6-cf80-474a-a582-ea693e5e1e16" path="/var/lib/kubelet/pods/43fd48e6-cf80-474a-a582-ea693e5e1e16/volumes" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.744402 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5205be66-cc23-4858-969e-b54efa4e6695" path="/var/lib/kubelet/pods/5205be66-cc23-4858-969e-b54efa4e6695/volumes" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.841262 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-pmjrt"] Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.980217 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" event={"ID":"11ffb893-0b90-4b17-9a48-a67f16eca6ac","Type":"ContainerDied","Data":"93d16dade0d2dd19fa9f7bbe1fcb7dd7ec137b77fb39bf1d62a978a7a4ce68bf"} Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.980273 4796 scope.go:117] "RemoveContainer" containerID="f1d58ca672e086cc6776ef863690765aa30e07a35ac3fe61ce4503a0900d9e0b" Sep 30 16:26:56 crc kubenswrapper[4796]: I0930 16:26:56.980291 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-hpsg7" Sep 30 16:26:57 crc kubenswrapper[4796]: I0930 16:26:57.020621 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-hpsg7"] Sep 30 16:26:57 crc kubenswrapper[4796]: I0930 16:26:57.029289 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-hpsg7"] Sep 30 16:26:57 crc kubenswrapper[4796]: I0930 16:26:57.186192 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kq8hq"] Sep 30 16:26:57 crc kubenswrapper[4796]: W0930 16:26:57.771660 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5bcbf15_75a7_4b1f_bb3d_4d58f59de746.slice/crio-a2790f6f9076f370412ac5e0e19fdf560c810cc98cbe26f66cf4ec9a07e405b2 WatchSource:0}: Error finding container a2790f6f9076f370412ac5e0e19fdf560c810cc98cbe26f66cf4ec9a07e405b2: Status 404 returned error can't find the container with id a2790f6f9076f370412ac5e0e19fdf560c810cc98cbe26f66cf4ec9a07e405b2 Sep 30 16:26:57 crc kubenswrapper[4796]: I0930 16:26:57.991490 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-pmjrt" event={"ID":"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746","Type":"ContainerStarted","Data":"a2790f6f9076f370412ac5e0e19fdf560c810cc98cbe26f66cf4ec9a07e405b2"} Sep 30 16:26:58 crc kubenswrapper[4796]: I0930 16:26:58.743553 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ffb893-0b90-4b17-9a48-a67f16eca6ac" path="/var/lib/kubelet/pods/11ffb893-0b90-4b17-9a48-a67f16eca6ac/volumes" Sep 30 16:26:59 crc kubenswrapper[4796]: W0930 16:26:59.521078 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11b92058_788c_40df_a406_afebc51358ca.slice/crio-41366ada4ad7e963ec1af424e079d881c7bcbf017efe23fc1bfff98f8c284341 WatchSource:0}: Error finding container 41366ada4ad7e963ec1af424e079d881c7bcbf017efe23fc1bfff98f8c284341: Status 404 returned error can't find the container with id 41366ada4ad7e963ec1af424e079d881c7bcbf017efe23fc1bfff98f8c284341 Sep 30 16:27:00 crc kubenswrapper[4796]: I0930 16:27:00.003023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" event={"ID":"11b92058-788c-40df-a406-afebc51358ca","Type":"ContainerStarted","Data":"41366ada4ad7e963ec1af424e079d881c7bcbf017efe23fc1bfff98f8c284341"} Sep 30 16:27:01 crc kubenswrapper[4796]: I0930 16:27:01.100100 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:27:01 crc kubenswrapper[4796]: I0930 16:27:01.100154 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:27:04 crc kubenswrapper[4796]: I0930 16:27:04.892747 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 16:27:05 crc kubenswrapper[4796]: I0930 16:27:05.065802 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a121b29f-66c2-4b9f-99b2-9a6c4a46473f","Type":"ContainerStarted","Data":"abb813ae7cabb7d3df999aa06b415f69d4d2250455ed263c32cd9dd012dec401"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.077509 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2bb8f271-8840-42e7-984c-31cf8fc8bea0","Type":"ContainerStarted","Data":"52b5e91e17521f3c402670a076ff07c9df28e0fb75a45a9d2b48423ae29980d8"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.084868 4796 generic.go:334] "Generic (PLEG): container finished" podID="11b92058-788c-40df-a406-afebc51358ca" containerID="8730bbebc1c3d4029c8c09cd2c745527f425996209fabb1d61c59cc744923acc" exitCode=0 Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.085213 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" event={"ID":"11b92058-788c-40df-a406-afebc51358ca","Type":"ContainerDied","Data":"8730bbebc1c3d4029c8c09cd2c745527f425996209fabb1d61c59cc744923acc"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.089404 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" event={"ID":"e6909bbc-1c68-43ff-afd5-6fbb1078297c","Type":"ContainerStarted","Data":"79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.090179 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.096579 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee37a44-c6b6-4d75-ad41-9ff1402925f1","Type":"ContainerStarted","Data":"677365ad12205b095e7ab1f45f8893d7344d64c584f08a009446e2d97c40ed89"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.107377 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"85b72cb4-9a34-4d07-8b11-dc60f4d540c9","Type":"ContainerStarted","Data":"5ef322fdabccd36131b0f7ee72ce0801594b9222998816f53d64776e6eb817e5"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.111603 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"94a4df46-32b9-4a28-848e-4c90a2660b82","Type":"ContainerStarted","Data":"5c329102bb30d66d6f487b5d3bcd280969be8252e414457d23e9e33c0abd1f03"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.112213 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.114161 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73","Type":"ContainerStarted","Data":"8a60af1eaba786e386562dfc364d8f6a95c91b30228085291d68c87f3723546f"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.115934 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6162abb2-dddc-4cb6-bd50-0e71ae9534bd","Type":"ContainerStarted","Data":"aea883e2889dd44757b6c081362ae32ccfbe7d9ce262e11080fd132ff5b4af65"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.133864 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vfdx6" event={"ID":"713f98aa-c3ec-4627-bc90-a1a577ca1c27","Type":"ContainerStarted","Data":"834e1791b18ea43fe954c286f657e46b041109f3fd61bf46b179df67349f5576"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.134622 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vfdx6" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.146264 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lg6m7" event={"ID":"403ec30e-d599-4940-8eaf-6cccb3200ed9","Type":"ContainerStarted","Data":"410c293990adf17a337d9a73a53555fe899c871462fbf0ad8920891d43d1d11a"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.148636 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-pmjrt" event={"ID":"f5bcbf15-75a7-4b1f-bb3d-4d58f59de746","Type":"ContainerStarted","Data":"6254135c3978f55e35900607c9f9bbde2ef5b3df1534ce6a4b7adcc047a046cb"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.151567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f340e3c8-c819-488b-808a-f610a2b60ef7","Type":"ContainerStarted","Data":"34a40a88436b6e90088af035949673031b8e24b425677f98cb88ecb849690d1c"} Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.152269 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.186340 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" podStartSLOduration=23.707900322 podStartE2EDuration="24.186317479s" podCreationTimestamp="2025-09-30 16:26:42 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.798709993 +0000 UTC m=+906.811988520" lastFinishedPulling="2025-09-30 16:26:55.27712715 +0000 UTC m=+907.290405677" observedRunningTime="2025-09-30 16:27:06.177337366 +0000 UTC m=+918.190615903" watchObservedRunningTime="2025-09-30 16:27:06.186317479 +0000 UTC m=+918.199596016" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.249531 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-pmjrt" podStartSLOduration=3.709072106 podStartE2EDuration="11.249514231s" podCreationTimestamp="2025-09-30 16:26:55 +0000 UTC" firstStartedPulling="2025-09-30 16:26:57.783642581 +0000 UTC m=+909.796921098" lastFinishedPulling="2025-09-30 16:27:05.324084666 +0000 UTC m=+917.337363223" observedRunningTime="2025-09-30 16:27:06.244088462 +0000 UTC m=+918.257366989" watchObservedRunningTime="2025-09-30 16:27:06.249514231 +0000 UTC m=+918.262792758" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.301798 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=7.874284567 podStartE2EDuration="18.301777873s" podCreationTimestamp="2025-09-30 16:26:48 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.895729655 +0000 UTC m=+906.909008182" lastFinishedPulling="2025-09-30 16:27:05.323222961 +0000 UTC m=+917.336501488" observedRunningTime="2025-09-30 16:27:06.288566815 +0000 UTC m=+918.301845342" watchObservedRunningTime="2025-09-30 16:27:06.301777873 +0000 UTC m=+918.315056400" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.324508 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.42691224 podStartE2EDuration="20.324485908s" podCreationTimestamp="2025-09-30 16:26:46 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.799870667 +0000 UTC m=+906.813149194" lastFinishedPulling="2025-09-30 16:27:02.697444335 +0000 UTC m=+914.710722862" observedRunningTime="2025-09-30 16:27:06.318576445 +0000 UTC m=+918.331855082" watchObservedRunningTime="2025-09-30 16:27:06.324485908 +0000 UTC m=+918.337764435" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.350351 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vfdx6" podStartSLOduration=5.028279929 podStartE2EDuration="15.350330585s" podCreationTimestamp="2025-09-30 16:26:51 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.896102956 +0000 UTC m=+906.909381483" lastFinishedPulling="2025-09-30 16:27:05.218153612 +0000 UTC m=+917.231432139" observedRunningTime="2025-09-30 16:27:06.341890278 +0000 UTC m=+918.355168805" watchObservedRunningTime="2025-09-30 16:27:06.350330585 +0000 UTC m=+918.363609112" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.540117 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bsdtj"] Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.566508 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z7lhm"] Sep 30 16:27:06 crc kubenswrapper[4796]: E0930 16:27:06.566841 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ffb893-0b90-4b17-9a48-a67f16eca6ac" containerName="init" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.566854 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ffb893-0b90-4b17-9a48-a67f16eca6ac" containerName="init" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.567411 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ffb893-0b90-4b17-9a48-a67f16eca6ac" containerName="init" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.568278 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.569963 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.582549 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z7lhm"] Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.605588 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-config\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.605680 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.605725 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.605756 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.605817 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gtr6\" (UniqueName: \"kubernetes.io/projected/9d3537c0-52ed-4021-9d62-7f1027ae3f17-kube-api-access-7gtr6\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.706799 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-config\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.707162 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.707213 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.707258 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.707291 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gtr6\" (UniqueName: \"kubernetes.io/projected/9d3537c0-52ed-4021-9d62-7f1027ae3f17-kube-api-access-7gtr6\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.707945 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-config\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.708265 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.708845 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.708872 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.724400 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gtr6\" (UniqueName: \"kubernetes.io/projected/9d3537c0-52ed-4021-9d62-7f1027ae3f17-kube-api-access-7gtr6\") pod \"dnsmasq-dns-86db49b7ff-z7lhm\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:06 crc kubenswrapper[4796]: I0930 16:27:06.884543 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.130260 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z7lhm"] Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.164835 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" event={"ID":"11b92058-788c-40df-a406-afebc51358ca","Type":"ContainerStarted","Data":"65984ad7c085611548307dba3be32692dc3dd06bfca53aeab4de5333bc3811c2"} Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.168152 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"85b72cb4-9a34-4d07-8b11-dc60f4d540c9","Type":"ContainerStarted","Data":"cc69bf623d7584274c342b5a789333c0c7961d746496a432d471c22d7f868ba5"} Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.183400 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" podStartSLOduration=12.183383034 podStartE2EDuration="12.183383034s" podCreationTimestamp="2025-09-30 16:26:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:27:07.1825837 +0000 UTC m=+919.195862247" watchObservedRunningTime="2025-09-30 16:27:07.183383034 +0000 UTC m=+919.196661561" Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.184159 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a121b29f-66c2-4b9f-99b2-9a6c4a46473f","Type":"ContainerStarted","Data":"01e1c86d5dab9e8708b0370e7e0005af4a7cf37d0ccebe01261949161666a8e4"} Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.184200 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a121b29f-66c2-4b9f-99b2-9a6c4a46473f","Type":"ContainerStarted","Data":"86e2d23683af5e66787de8655a3047e5db7e1151786b7d45922790d24b06e05a"} Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.185810 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" event={"ID":"9d3537c0-52ed-4021-9d62-7f1027ae3f17","Type":"ContainerStarted","Data":"a4a0e3b5f42b5e8cb8039f9c3f955adf9516ce30b5da1a93943ff018db50bc61"} Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.189813 4796 generic.go:334] "Generic (PLEG): container finished" podID="403ec30e-d599-4940-8eaf-6cccb3200ed9" containerID="410c293990adf17a337d9a73a53555fe899c871462fbf0ad8920891d43d1d11a" exitCode=0 Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.190838 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lg6m7" event={"ID":"403ec30e-d599-4940-8eaf-6cccb3200ed9","Type":"ContainerDied","Data":"410c293990adf17a337d9a73a53555fe899c871462fbf0ad8920891d43d1d11a"} Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.208495 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.91122191 podStartE2EDuration="15.208470159s" podCreationTimestamp="2025-09-30 16:26:52 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.988212895 +0000 UTC m=+907.001491412" lastFinishedPulling="2025-09-30 16:27:05.285461134 +0000 UTC m=+917.298739661" observedRunningTime="2025-09-30 16:27:07.200635389 +0000 UTC m=+919.213913916" watchObservedRunningTime="2025-09-30 16:27:07.208470159 +0000 UTC m=+919.221748706" Sep 30 16:27:07 crc kubenswrapper[4796]: I0930 16:27:07.226841 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.227115414 podStartE2EDuration="12.226815666s" podCreationTimestamp="2025-09-30 16:26:55 +0000 UTC" firstStartedPulling="2025-09-30 16:27:05.057439933 +0000 UTC m=+917.070718460" lastFinishedPulling="2025-09-30 16:27:06.057140185 +0000 UTC m=+918.070418712" observedRunningTime="2025-09-30 16:27:07.219569874 +0000 UTC m=+919.232848411" watchObservedRunningTime="2025-09-30 16:27:07.226815666 +0000 UTC m=+919.240094193" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.199566 4796 generic.go:334] "Generic (PLEG): container finished" podID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerID="e124763efe374d1bbde6afc28d9f5fa5afe34f659f83cad82cbf5c6326344897" exitCode=0 Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.199635 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" event={"ID":"9d3537c0-52ed-4021-9d62-7f1027ae3f17","Type":"ContainerDied","Data":"e124763efe374d1bbde6afc28d9f5fa5afe34f659f83cad82cbf5c6326344897"} Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.203302 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lg6m7" event={"ID":"403ec30e-d599-4940-8eaf-6cccb3200ed9","Type":"ContainerStarted","Data":"9556c24fc869e6e4a69ed5863ce0a0405c89f93f0e2702397371f9c2d4813e56"} Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.204429 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.204482 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lg6m7" event={"ID":"403ec30e-d599-4940-8eaf-6cccb3200ed9","Type":"ContainerStarted","Data":"980469c837cc9962002d5d5f1b2467fc004f9093adffaa14765b5eaeca9314ba"} Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.204514 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.204544 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.204756 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerName="dnsmasq-dns" containerID="cri-o://79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f" gracePeriod=10 Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.255200 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-lg6m7" podStartSLOduration=7.95048286 podStartE2EDuration="17.255179647s" podCreationTimestamp="2025-09-30 16:26:51 +0000 UTC" firstStartedPulling="2025-09-30 16:26:55.078190641 +0000 UTC m=+907.091469188" lastFinishedPulling="2025-09-30 16:27:04.382887448 +0000 UTC m=+916.396165975" observedRunningTime="2025-09-30 16:27:08.248650725 +0000 UTC m=+920.261929252" watchObservedRunningTime="2025-09-30 16:27:08.255179647 +0000 UTC m=+920.268458174" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.631968 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.748010 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-config\") pod \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.748304 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-dns-svc\") pod \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.748450 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvkqb\" (UniqueName: \"kubernetes.io/projected/e6909bbc-1c68-43ff-afd5-6fbb1078297c-kube-api-access-rvkqb\") pod \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\" (UID: \"e6909bbc-1c68-43ff-afd5-6fbb1078297c\") " Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.753551 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6909bbc-1c68-43ff-afd5-6fbb1078297c-kube-api-access-rvkqb" (OuterVolumeSpecName: "kube-api-access-rvkqb") pod "e6909bbc-1c68-43ff-afd5-6fbb1078297c" (UID: "e6909bbc-1c68-43ff-afd5-6fbb1078297c"). InnerVolumeSpecName "kube-api-access-rvkqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.781874 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.799449 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-config" (OuterVolumeSpecName: "config") pod "e6909bbc-1c68-43ff-afd5-6fbb1078297c" (UID: "e6909bbc-1c68-43ff-afd5-6fbb1078297c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.803405 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6909bbc-1c68-43ff-afd5-6fbb1078297c" (UID: "e6909bbc-1c68-43ff-afd5-6fbb1078297c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.851282 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.851328 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6909bbc-1c68-43ff-afd5-6fbb1078297c-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.851340 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvkqb\" (UniqueName: \"kubernetes.io/projected/e6909bbc-1c68-43ff-afd5-6fbb1078297c-kube-api-access-rvkqb\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.907719 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.907776 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 30 16:27:08 crc kubenswrapper[4796]: I0930 16:27:08.955067 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.215247 4796 generic.go:334] "Generic (PLEG): container finished" podID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerID="79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f" exitCode=0 Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.215341 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.215329 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" event={"ID":"e6909bbc-1c68-43ff-afd5-6fbb1078297c","Type":"ContainerDied","Data":"79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f"} Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.215483 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bsdtj" event={"ID":"e6909bbc-1c68-43ff-afd5-6fbb1078297c","Type":"ContainerDied","Data":"ac041d45eecc4be46f0a04fe57f1761108c4292bbd35e09ddb6c93961dd64af7"} Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.215519 4796 scope.go:117] "RemoveContainer" containerID="79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.217442 4796 generic.go:334] "Generic (PLEG): container finished" podID="6162abb2-dddc-4cb6-bd50-0e71ae9534bd" containerID="aea883e2889dd44757b6c081362ae32ccfbe7d9ce262e11080fd132ff5b4af65" exitCode=0 Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.217479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6162abb2-dddc-4cb6-bd50-0e71ae9534bd","Type":"ContainerDied","Data":"aea883e2889dd44757b6c081362ae32ccfbe7d9ce262e11080fd132ff5b4af65"} Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.224152 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" event={"ID":"9d3537c0-52ed-4021-9d62-7f1027ae3f17","Type":"ContainerStarted","Data":"8f5810f1a3005e3916c0b71ad9dcde96bf6391791c015e3321790becfeb2de13"} Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.225091 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.290044 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" podStartSLOduration=3.290016038 podStartE2EDuration="3.290016038s" podCreationTimestamp="2025-09-30 16:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:27:09.280389016 +0000 UTC m=+921.293667553" watchObservedRunningTime="2025-09-30 16:27:09.290016038 +0000 UTC m=+921.303294595" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.428302 4796 scope.go:117] "RemoveContainer" containerID="f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.449213 4796 scope.go:117] "RemoveContainer" containerID="79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.449401 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bsdtj"] Sep 30 16:27:09 crc kubenswrapper[4796]: E0930 16:27:09.449867 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f\": container with ID starting with 79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f not found: ID does not exist" containerID="79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.449914 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f"} err="failed to get container status \"79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f\": rpc error: code = NotFound desc = could not find container \"79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f\": container with ID starting with 79b56fd1d105983b345e652cfe03e5e31cfbc18c91c6e65c745419a4be3bca9f not found: ID does not exist" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.449949 4796 scope.go:117] "RemoveContainer" containerID="f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb" Sep 30 16:27:09 crc kubenswrapper[4796]: E0930 16:27:09.450340 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb\": container with ID starting with f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb not found: ID does not exist" containerID="f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.450377 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb"} err="failed to get container status \"f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb\": rpc error: code = NotFound desc = could not find container \"f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb\": container with ID starting with f4163704eaae5f2bf8c7f6d66120000d01127ebbb18f55956b9495a5d64215eb not found: ID does not exist" Sep 30 16:27:09 crc kubenswrapper[4796]: I0930 16:27:09.451143 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bsdtj"] Sep 30 16:27:10 crc kubenswrapper[4796]: I0930 16:27:10.239024 4796 generic.go:334] "Generic (PLEG): container finished" podID="2bb8f271-8840-42e7-984c-31cf8fc8bea0" containerID="52b5e91e17521f3c402670a076ff07c9df28e0fb75a45a9d2b48423ae29980d8" exitCode=0 Sep 30 16:27:10 crc kubenswrapper[4796]: I0930 16:27:10.239111 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2bb8f271-8840-42e7-984c-31cf8fc8bea0","Type":"ContainerDied","Data":"52b5e91e17521f3c402670a076ff07c9df28e0fb75a45a9d2b48423ae29980d8"} Sep 30 16:27:10 crc kubenswrapper[4796]: I0930 16:27:10.246573 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6162abb2-dddc-4cb6-bd50-0e71ae9534bd","Type":"ContainerStarted","Data":"efa35d541a46514dba28ad1152c7eff01b16521e04003a54c57709196fd73cdf"} Sep 30 16:27:10 crc kubenswrapper[4796]: I0930 16:27:10.299189 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.699890168 podStartE2EDuration="26.299167476s" podCreationTimestamp="2025-09-30 16:26:44 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.799673381 +0000 UTC m=+906.812951898" lastFinishedPulling="2025-09-30 16:27:04.398950659 +0000 UTC m=+916.412229206" observedRunningTime="2025-09-30 16:27:10.294440067 +0000 UTC m=+922.307718624" watchObservedRunningTime="2025-09-30 16:27:10.299167476 +0000 UTC m=+922.312446013" Sep 30 16:27:10 crc kubenswrapper[4796]: I0930 16:27:10.306226 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 30 16:27:10 crc kubenswrapper[4796]: I0930 16:27:10.742462 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" path="/var/lib/kubelet/pods/e6909bbc-1c68-43ff-afd5-6fbb1078297c/volumes" Sep 30 16:27:11 crc kubenswrapper[4796]: I0930 16:27:11.106726 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:27:11 crc kubenswrapper[4796]: I0930 16:27:11.256717 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2bb8f271-8840-42e7-984c-31cf8fc8bea0","Type":"ContainerStarted","Data":"02c3f72152a661b9cb5b6fff355caa25699fece1169964ceab8c9d51a5576945"} Sep 30 16:27:11 crc kubenswrapper[4796]: I0930 16:27:11.715063 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 30 16:27:11 crc kubenswrapper[4796]: I0930 16:27:11.738713 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 30 16:27:11 crc kubenswrapper[4796]: I0930 16:27:11.740512 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=16.349719596 podStartE2EDuration="26.740492696s" podCreationTimestamp="2025-09-30 16:26:45 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.888330088 +0000 UTC m=+906.901608615" lastFinishedPulling="2025-09-30 16:27:05.279103188 +0000 UTC m=+917.292381715" observedRunningTime="2025-09-30 16:27:11.283837707 +0000 UTC m=+923.297116234" watchObservedRunningTime="2025-09-30 16:27:11.740492696 +0000 UTC m=+923.753771233" Sep 30 16:27:11 crc kubenswrapper[4796]: I0930 16:27:11.790008 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.302756 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.603225 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 30 16:27:12 crc kubenswrapper[4796]: E0930 16:27:12.603872 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerName="dnsmasq-dns" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.603890 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerName="dnsmasq-dns" Sep 30 16:27:12 crc kubenswrapper[4796]: E0930 16:27:12.603913 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerName="init" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.603921 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerName="init" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.604151 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6909bbc-1c68-43ff-afd5-6fbb1078297c" containerName="dnsmasq-dns" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.605184 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.608216 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.608266 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vbj98" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.608216 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.609414 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.620945 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623161 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623212 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43d2a8c3-5a3c-475d-a348-843d5f59b008-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623236 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623380 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxgmn\" (UniqueName: \"kubernetes.io/projected/43d2a8c3-5a3c-475d-a348-843d5f59b008-kube-api-access-lxgmn\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623487 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43d2a8c3-5a3c-475d-a348-843d5f59b008-scripts\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d2a8c3-5a3c-475d-a348-843d5f59b008-config\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.623798 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725171 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725213 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43d2a8c3-5a3c-475d-a348-843d5f59b008-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725231 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725266 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxgmn\" (UniqueName: \"kubernetes.io/projected/43d2a8c3-5a3c-475d-a348-843d5f59b008-kube-api-access-lxgmn\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725299 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43d2a8c3-5a3c-475d-a348-843d5f59b008-scripts\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725364 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d2a8c3-5a3c-475d-a348-843d5f59b008-config\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.725383 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.726606 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43d2a8c3-5a3c-475d-a348-843d5f59b008-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.726999 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d2a8c3-5a3c-475d-a348-843d5f59b008-config\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.727286 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43d2a8c3-5a3c-475d-a348-843d5f59b008-scripts\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.731945 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.733134 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.741338 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/43d2a8c3-5a3c-475d-a348-843d5f59b008-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.750883 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxgmn\" (UniqueName: \"kubernetes.io/projected/43d2a8c3-5a3c-475d-a348-843d5f59b008-kube-api-access-lxgmn\") pod \"ovn-northd-0\" (UID: \"43d2a8c3-5a3c-475d-a348-843d5f59b008\") " pod="openstack/ovn-northd-0" Sep 30 16:27:12 crc kubenswrapper[4796]: I0930 16:27:12.926687 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 16:27:13 crc kubenswrapper[4796]: I0930 16:27:13.355585 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 16:27:14 crc kubenswrapper[4796]: I0930 16:27:14.282452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"43d2a8c3-5a3c-475d-a348-843d5f59b008","Type":"ContainerStarted","Data":"881f0bf2a1e749df0f675cc1fbdd806ec9715d235f13c4f5a55237b3ed041c09"} Sep 30 16:27:15 crc kubenswrapper[4796]: I0930 16:27:15.295685 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"43d2a8c3-5a3c-475d-a348-843d5f59b008","Type":"ContainerStarted","Data":"38748085e9e092c36dd5232befde80d6f2efd67881a4a99b1f90f338605d46a0"} Sep 30 16:27:15 crc kubenswrapper[4796]: I0930 16:27:15.296506 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 30 16:27:15 crc kubenswrapper[4796]: I0930 16:27:15.296540 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"43d2a8c3-5a3c-475d-a348-843d5f59b008","Type":"ContainerStarted","Data":"a53d3ee6dbd93544015c4d48e102a1fc0946b51c9ed3294878e5f28044bf063b"} Sep 30 16:27:15 crc kubenswrapper[4796]: I0930 16:27:15.324506 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.2699298 podStartE2EDuration="3.324487878s" podCreationTimestamp="2025-09-30 16:27:12 +0000 UTC" firstStartedPulling="2025-09-30 16:27:13.36754779 +0000 UTC m=+925.380826317" lastFinishedPulling="2025-09-30 16:27:14.422105858 +0000 UTC m=+926.435384395" observedRunningTime="2025-09-30 16:27:15.321581312 +0000 UTC m=+927.334859849" watchObservedRunningTime="2025-09-30 16:27:15.324487878 +0000 UTC m=+927.337766415" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.324041 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.324101 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.393085 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.452560 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.453014 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.507854 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.886200 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.940919 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kq8hq"] Sep 30 16:27:16 crc kubenswrapper[4796]: I0930 16:27:16.941569 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="dnsmasq-dns" containerID="cri-o://65984ad7c085611548307dba3be32692dc3dd06bfca53aeab4de5333bc3811c2" gracePeriod=10 Sep 30 16:27:17 crc kubenswrapper[4796]: I0930 16:27:17.319456 4796 generic.go:334] "Generic (PLEG): container finished" podID="11b92058-788c-40df-a406-afebc51358ca" containerID="65984ad7c085611548307dba3be32692dc3dd06bfca53aeab4de5333bc3811c2" exitCode=0 Sep 30 16:27:17 crc kubenswrapper[4796]: I0930 16:27:17.319747 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" event={"ID":"11b92058-788c-40df-a406-afebc51358ca","Type":"ContainerDied","Data":"65984ad7c085611548307dba3be32692dc3dd06bfca53aeab4de5333bc3811c2"} Sep 30 16:27:17 crc kubenswrapper[4796]: I0930 16:27:17.391898 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 30 16:27:17 crc kubenswrapper[4796]: I0930 16:27:17.391995 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.424703 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.588258 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-sc9nt"] Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.594935 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.624192 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sc9nt"] Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.730773 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-dns-svc\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.730859 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-config\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.731012 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw5bp\" (UniqueName: \"kubernetes.io/projected/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-kube-api-access-zw5bp\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.731055 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.731170 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.833383 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-dns-svc\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.833476 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-config\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.833556 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw5bp\" (UniqueName: \"kubernetes.io/projected/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-kube-api-access-zw5bp\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.833588 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.833644 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.835352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-config\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.835651 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.835700 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.836077 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-dns-svc\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.863642 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw5bp\" (UniqueName: \"kubernetes.io/projected/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-kube-api-access-zw5bp\") pod \"dnsmasq-dns-698758b865-sc9nt\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:18 crc kubenswrapper[4796]: I0930 16:27:18.918560 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.338234 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sc9nt"] Sep 30 16:27:19 crc kubenswrapper[4796]: W0930 16:27:19.345258 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9d104f1_e734_4c39_9e74_f7b39cd9dbde.slice/crio-d0a4357cf73b9a8c89941e2a80cf142b8b696ed3bc981d45e0edd9ea22ac222e WatchSource:0}: Error finding container d0a4357cf73b9a8c89941e2a80cf142b8b696ed3bc981d45e0edd9ea22ac222e: Status 404 returned error can't find the container with id d0a4357cf73b9a8c89941e2a80cf142b8b696ed3bc981d45e0edd9ea22ac222e Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.656106 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.665703 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.668844 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.669156 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-qhmwn" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.669319 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.679553 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.701477 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.755573 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/24528035-6f6d-4269-9e88-6171795db8a7-cache\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.755655 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv8g9\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-kube-api-access-kv8g9\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.756616 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.756861 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.756908 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/24528035-6f6d-4269-9e88-6171795db8a7-lock\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.859074 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.859601 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.860438 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.860536 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/24528035-6f6d-4269-9e88-6171795db8a7-lock\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.860686 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/24528035-6f6d-4269-9e88-6171795db8a7-cache\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.860754 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv8g9\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-kube-api-access-kv8g9\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: E0930 16:27:19.860834 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 16:27:19 crc kubenswrapper[4796]: E0930 16:27:19.860865 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 16:27:19 crc kubenswrapper[4796]: E0930 16:27:19.860932 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift podName:24528035-6f6d-4269-9e88-6171795db8a7 nodeName:}" failed. No retries permitted until 2025-09-30 16:27:20.360904665 +0000 UTC m=+932.374183302 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift") pod "swift-storage-0" (UID: "24528035-6f6d-4269-9e88-6171795db8a7") : configmap "swift-ring-files" not found Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.861706 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/24528035-6f6d-4269-9e88-6171795db8a7-lock\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.862039 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/24528035-6f6d-4269-9e88-6171795db8a7-cache\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.881587 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:19 crc kubenswrapper[4796]: I0930 16:27:19.894135 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv8g9\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-kube-api-access-kv8g9\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:20 crc kubenswrapper[4796]: I0930 16:27:20.352461 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sc9nt" event={"ID":"f9d104f1-e734-4c39-9e74-f7b39cd9dbde","Type":"ContainerStarted","Data":"d0a4357cf73b9a8c89941e2a80cf142b8b696ed3bc981d45e0edd9ea22ac222e"} Sep 30 16:27:20 crc kubenswrapper[4796]: I0930 16:27:20.369632 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:20 crc kubenswrapper[4796]: E0930 16:27:20.369866 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 16:27:20 crc kubenswrapper[4796]: E0930 16:27:20.369894 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 16:27:20 crc kubenswrapper[4796]: E0930 16:27:20.369951 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift podName:24528035-6f6d-4269-9e88-6171795db8a7 nodeName:}" failed. No retries permitted until 2025-09-30 16:27:21.369933959 +0000 UTC m=+933.383212486 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift") pod "swift-storage-0" (UID: "24528035-6f6d-4269-9e88-6171795db8a7") : configmap "swift-ring-files" not found Sep 30 16:27:21 crc kubenswrapper[4796]: I0930 16:27:21.104911 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: connect: connection refused" Sep 30 16:27:21 crc kubenswrapper[4796]: I0930 16:27:21.387230 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:21 crc kubenswrapper[4796]: E0930 16:27:21.388067 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 16:27:21 crc kubenswrapper[4796]: E0930 16:27:21.388270 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 16:27:21 crc kubenswrapper[4796]: E0930 16:27:21.388474 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift podName:24528035-6f6d-4269-9e88-6171795db8a7 nodeName:}" failed. No retries permitted until 2025-09-30 16:27:23.388448062 +0000 UTC m=+935.401726619 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift") pod "swift-storage-0" (UID: "24528035-6f6d-4269-9e88-6171795db8a7") : configmap "swift-ring-files" not found Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.074274 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-c8s6r"] Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.076087 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.095085 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c8s6r"] Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.212306 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4k57\" (UniqueName: \"kubernetes.io/projected/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b-kube-api-access-m4k57\") pod \"glance-db-create-c8s6r\" (UID: \"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b\") " pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.314104 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4k57\" (UniqueName: \"kubernetes.io/projected/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b-kube-api-access-m4k57\") pod \"glance-db-create-c8s6r\" (UID: \"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b\") " pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.343626 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4k57\" (UniqueName: \"kubernetes.io/projected/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b-kube-api-access-m4k57\") pod \"glance-db-create-c8s6r\" (UID: \"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b\") " pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.373055 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" event={"ID":"11b92058-788c-40df-a406-afebc51358ca","Type":"ContainerDied","Data":"41366ada4ad7e963ec1af424e079d881c7bcbf017efe23fc1bfff98f8c284341"} Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.373098 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41366ada4ad7e963ec1af424e079d881c7bcbf017efe23fc1bfff98f8c284341" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.375782 4796 generic.go:334] "Generic (PLEG): container finished" podID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerID="1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19" exitCode=0 Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.375827 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sc9nt" event={"ID":"f9d104f1-e734-4c39-9e74-f7b39cd9dbde","Type":"ContainerDied","Data":"1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19"} Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.402136 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.544880 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.618639 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-ovsdbserver-nb\") pod \"11b92058-788c-40df-a406-afebc51358ca\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.618779 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-config\") pod \"11b92058-788c-40df-a406-afebc51358ca\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.618856 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-dns-svc\") pod \"11b92058-788c-40df-a406-afebc51358ca\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.618888 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncgzk\" (UniqueName: \"kubernetes.io/projected/11b92058-788c-40df-a406-afebc51358ca-kube-api-access-ncgzk\") pod \"11b92058-788c-40df-a406-afebc51358ca\" (UID: \"11b92058-788c-40df-a406-afebc51358ca\") " Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.623801 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b92058-788c-40df-a406-afebc51358ca-kube-api-access-ncgzk" (OuterVolumeSpecName: "kube-api-access-ncgzk") pod "11b92058-788c-40df-a406-afebc51358ca" (UID: "11b92058-788c-40df-a406-afebc51358ca"). InnerVolumeSpecName "kube-api-access-ncgzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.667591 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-config" (OuterVolumeSpecName: "config") pod "11b92058-788c-40df-a406-afebc51358ca" (UID: "11b92058-788c-40df-a406-afebc51358ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.672633 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11b92058-788c-40df-a406-afebc51358ca" (UID: "11b92058-788c-40df-a406-afebc51358ca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.677076 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11b92058-788c-40df-a406-afebc51358ca" (UID: "11b92058-788c-40df-a406-afebc51358ca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.722536 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.722893 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.722910 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b92058-788c-40df-a406-afebc51358ca-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.722921 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncgzk\" (UniqueName: \"kubernetes.io/projected/11b92058-788c-40df-a406-afebc51358ca-kube-api-access-ncgzk\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:22 crc kubenswrapper[4796]: I0930 16:27:22.891742 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c8s6r"] Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.387568 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sc9nt" event={"ID":"f9d104f1-e734-4c39-9e74-f7b39cd9dbde","Type":"ContainerStarted","Data":"5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae"} Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.389578 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.392000 4796 generic.go:334] "Generic (PLEG): container finished" podID="b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b" containerID="b3e0c4f6ec320b26192f04c26c7e6b2375db967b7d2e999c175ca55bba670a50" exitCode=0 Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.392093 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kq8hq" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.392575 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c8s6r" event={"ID":"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b","Type":"ContainerDied","Data":"b3e0c4f6ec320b26192f04c26c7e6b2375db967b7d2e999c175ca55bba670a50"} Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.392604 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c8s6r" event={"ID":"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b","Type":"ContainerStarted","Data":"a6a775b5643abc10dd0c68477a50aa5d69e3a08e4518c2843a931b36026bc231"} Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.420789 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-sc9nt" podStartSLOduration=5.420771259 podStartE2EDuration="5.420771259s" podCreationTimestamp="2025-09-30 16:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:27:23.41636046 +0000 UTC m=+935.429639017" watchObservedRunningTime="2025-09-30 16:27:23.420771259 +0000 UTC m=+935.434049786" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.434812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:23 crc kubenswrapper[4796]: E0930 16:27:23.435028 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 16:27:23 crc kubenswrapper[4796]: E0930 16:27:23.435049 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 16:27:23 crc kubenswrapper[4796]: E0930 16:27:23.435099 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift podName:24528035-6f6d-4269-9e88-6171795db8a7 nodeName:}" failed. No retries permitted until 2025-09-30 16:27:27.435081938 +0000 UTC m=+939.448360465 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift") pod "swift-storage-0" (UID: "24528035-6f6d-4269-9e88-6171795db8a7") : configmap "swift-ring-files" not found Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.459933 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kq8hq"] Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.470695 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kq8hq"] Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.592465 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-m6xqp"] Sep 30 16:27:23 crc kubenswrapper[4796]: E0930 16:27:23.593076 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="dnsmasq-dns" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.593110 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="dnsmasq-dns" Sep 30 16:27:23 crc kubenswrapper[4796]: E0930 16:27:23.593181 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="init" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.593195 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="init" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.593493 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b92058-788c-40df-a406-afebc51358ca" containerName="dnsmasq-dns" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.597276 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.600339 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.601478 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.608665 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.610436 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m6xqp"] Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740207 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgz2j\" (UniqueName: \"kubernetes.io/projected/420fc105-e180-4b89-af90-af130839e938-kube-api-access-qgz2j\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740258 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-scripts\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740375 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-ring-data-devices\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740403 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-swiftconf\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740427 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/420fc105-e180-4b89-af90-af130839e938-etc-swift\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740465 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-combined-ca-bundle\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.740491 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-dispersionconf\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.842554 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgz2j\" (UniqueName: \"kubernetes.io/projected/420fc105-e180-4b89-af90-af130839e938-kube-api-access-qgz2j\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.842645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-scripts\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.842842 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-ring-data-devices\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.842889 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-swiftconf\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.842943 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/420fc105-e180-4b89-af90-af130839e938-etc-swift\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.843054 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-combined-ca-bundle\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.843101 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-dispersionconf\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.843357 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/420fc105-e180-4b89-af90-af130839e938-etc-swift\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.844166 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-ring-data-devices\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.844314 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-scripts\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.852959 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-dispersionconf\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.853313 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-swiftconf\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.854284 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-combined-ca-bundle\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.871017 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgz2j\" (UniqueName: \"kubernetes.io/projected/420fc105-e180-4b89-af90-af130839e938-kube-api-access-qgz2j\") pod \"swift-ring-rebalance-m6xqp\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:23 crc kubenswrapper[4796]: I0930 16:27:23.938535 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:24 crc kubenswrapper[4796]: I0930 16:27:24.394762 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m6xqp"] Sep 30 16:27:24 crc kubenswrapper[4796]: W0930 16:27:24.396740 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod420fc105_e180_4b89_af90_af130839e938.slice/crio-630291bd6f4ef420f2f27a47428acdf3ed913463b6d691e5e277f1e842ee3c72 WatchSource:0}: Error finding container 630291bd6f4ef420f2f27a47428acdf3ed913463b6d691e5e277f1e842ee3c72: Status 404 returned error can't find the container with id 630291bd6f4ef420f2f27a47428acdf3ed913463b6d691e5e277f1e842ee3c72 Sep 30 16:27:24 crc kubenswrapper[4796]: I0930 16:27:24.747047 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b92058-788c-40df-a406-afebc51358ca" path="/var/lib/kubelet/pods/11b92058-788c-40df-a406-afebc51358ca/volumes" Sep 30 16:27:24 crc kubenswrapper[4796]: I0930 16:27:24.818587 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:24 crc kubenswrapper[4796]: I0930 16:27:24.970052 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4k57\" (UniqueName: \"kubernetes.io/projected/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b-kube-api-access-m4k57\") pod \"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b\" (UID: \"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b\") " Sep 30 16:27:24 crc kubenswrapper[4796]: I0930 16:27:24.981771 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b-kube-api-access-m4k57" (OuterVolumeSpecName: "kube-api-access-m4k57") pod "b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b" (UID: "b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b"). InnerVolumeSpecName "kube-api-access-m4k57". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:25 crc kubenswrapper[4796]: I0930 16:27:25.072424 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4k57\" (UniqueName: \"kubernetes.io/projected/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b-kube-api-access-m4k57\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:25 crc kubenswrapper[4796]: I0930 16:27:25.409093 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c8s6r" Sep 30 16:27:25 crc kubenswrapper[4796]: I0930 16:27:25.409104 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c8s6r" event={"ID":"b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b","Type":"ContainerDied","Data":"a6a775b5643abc10dd0c68477a50aa5d69e3a08e4518c2843a931b36026bc231"} Sep 30 16:27:25 crc kubenswrapper[4796]: I0930 16:27:25.409809 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6a775b5643abc10dd0c68477a50aa5d69e3a08e4518c2843a931b36026bc231" Sep 30 16:27:25 crc kubenswrapper[4796]: I0930 16:27:25.410851 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6xqp" event={"ID":"420fc105-e180-4b89-af90-af130839e938","Type":"ContainerStarted","Data":"630291bd6f4ef420f2f27a47428acdf3ed913463b6d691e5e277f1e842ee3c72"} Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.330279 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-l49w6"] Sep 30 16:27:26 crc kubenswrapper[4796]: E0930 16:27:26.330883 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b" containerName="mariadb-database-create" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.330908 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b" containerName="mariadb-database-create" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.331157 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b" containerName="mariadb-database-create" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.331982 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.335963 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-l49w6"] Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.398917 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7h44\" (UniqueName: \"kubernetes.io/projected/06de8b69-fc4c-48b6-8fd6-6862a82b5620-kube-api-access-j7h44\") pod \"keystone-db-create-l49w6\" (UID: \"06de8b69-fc4c-48b6-8fd6-6862a82b5620\") " pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.501735 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7h44\" (UniqueName: \"kubernetes.io/projected/06de8b69-fc4c-48b6-8fd6-6862a82b5620-kube-api-access-j7h44\") pod \"keystone-db-create-l49w6\" (UID: \"06de8b69-fc4c-48b6-8fd6-6862a82b5620\") " pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.522762 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7h44\" (UniqueName: \"kubernetes.io/projected/06de8b69-fc4c-48b6-8fd6-6862a82b5620-kube-api-access-j7h44\") pod \"keystone-db-create-l49w6\" (UID: \"06de8b69-fc4c-48b6-8fd6-6862a82b5620\") " pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.613767 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-q2wwg"] Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.615113 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.625862 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q2wwg"] Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.661559 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.715072 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq776\" (UniqueName: \"kubernetes.io/projected/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed-kube-api-access-jq776\") pod \"placement-db-create-q2wwg\" (UID: \"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed\") " pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.816849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq776\" (UniqueName: \"kubernetes.io/projected/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed-kube-api-access-jq776\") pod \"placement-db-create-q2wwg\" (UID: \"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed\") " pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.850938 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq776\" (UniqueName: \"kubernetes.io/projected/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed-kube-api-access-jq776\") pod \"placement-db-create-q2wwg\" (UID: \"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed\") " pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:26 crc kubenswrapper[4796]: I0930 16:27:26.942870 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:27 crc kubenswrapper[4796]: I0930 16:27:27.528040 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:27 crc kubenswrapper[4796]: E0930 16:27:27.528226 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 16:27:27 crc kubenswrapper[4796]: E0930 16:27:27.528498 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 16:27:27 crc kubenswrapper[4796]: E0930 16:27:27.528557 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift podName:24528035-6f6d-4269-9e88-6171795db8a7 nodeName:}" failed. No retries permitted until 2025-09-30 16:27:35.528537987 +0000 UTC m=+947.541816514 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift") pod "swift-storage-0" (UID: "24528035-6f6d-4269-9e88-6171795db8a7") : configmap "swift-ring-files" not found Sep 30 16:27:27 crc kubenswrapper[4796]: I0930 16:27:27.989084 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q2wwg"] Sep 30 16:27:27 crc kubenswrapper[4796]: W0930 16:27:27.989598 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcafd3c1f_17cb_426c_8dbe_f0f8dc0bc5ed.slice/crio-285b21ecb0dfdfc1aaa5a1f7c93a5d74a6b32530b9c583823f906018d6353cfc WatchSource:0}: Error finding container 285b21ecb0dfdfc1aaa5a1f7c93a5d74a6b32530b9c583823f906018d6353cfc: Status 404 returned error can't find the container with id 285b21ecb0dfdfc1aaa5a1f7c93a5d74a6b32530b9c583823f906018d6353cfc Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.011412 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.098892 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-l49w6"] Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.437285 4796 generic.go:334] "Generic (PLEG): container finished" podID="cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed" containerID="3c0cd8a3cc475176ea5ea6cd2f86893bb0cc893082c69e0999f53eb0bfa60e55" exitCode=0 Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.437362 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q2wwg" event={"ID":"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed","Type":"ContainerDied","Data":"3c0cd8a3cc475176ea5ea6cd2f86893bb0cc893082c69e0999f53eb0bfa60e55"} Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.437394 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q2wwg" event={"ID":"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed","Type":"ContainerStarted","Data":"285b21ecb0dfdfc1aaa5a1f7c93a5d74a6b32530b9c583823f906018d6353cfc"} Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.438821 4796 generic.go:334] "Generic (PLEG): container finished" podID="06de8b69-fc4c-48b6-8fd6-6862a82b5620" containerID="89d923d17b385bad1bcb3c5c70e5688be4bd3fe4a17652dc29efb54a8208e47c" exitCode=0 Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.438836 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l49w6" event={"ID":"06de8b69-fc4c-48b6-8fd6-6862a82b5620","Type":"ContainerDied","Data":"89d923d17b385bad1bcb3c5c70e5688be4bd3fe4a17652dc29efb54a8208e47c"} Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.438859 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l49w6" event={"ID":"06de8b69-fc4c-48b6-8fd6-6862a82b5620","Type":"ContainerStarted","Data":"bd2a40c98617bf675815e1c750d3f7e063146eba728e0434c76760f736adb0d8"} Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.440461 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6xqp" event={"ID":"420fc105-e180-4b89-af90-af130839e938","Type":"ContainerStarted","Data":"864b2c97e33185765c97cbaae5b9767ed575fcd7df71b1a7819a846d49f8236b"} Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.504461 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-m6xqp" podStartSLOduration=2.340717233 podStartE2EDuration="5.504418s" podCreationTimestamp="2025-09-30 16:27:23 +0000 UTC" firstStartedPulling="2025-09-30 16:27:24.399513176 +0000 UTC m=+936.412791733" lastFinishedPulling="2025-09-30 16:27:27.563213973 +0000 UTC m=+939.576492500" observedRunningTime="2025-09-30 16:27:28.499851917 +0000 UTC m=+940.513130464" watchObservedRunningTime="2025-09-30 16:27:28.504418 +0000 UTC m=+940.517696567" Sep 30 16:27:28 crc kubenswrapper[4796]: I0930 16:27:28.920260 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.004863 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z7lhm"] Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.005210 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerName="dnsmasq-dns" containerID="cri-o://8f5810f1a3005e3916c0b71ad9dcde96bf6391791c015e3321790becfeb2de13" gracePeriod=10 Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.451766 4796 generic.go:334] "Generic (PLEG): container finished" podID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerID="8f5810f1a3005e3916c0b71ad9dcde96bf6391791c015e3321790becfeb2de13" exitCode=0 Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.451906 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" event={"ID":"9d3537c0-52ed-4021-9d62-7f1027ae3f17","Type":"ContainerDied","Data":"8f5810f1a3005e3916c0b71ad9dcde96bf6391791c015e3321790becfeb2de13"} Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.452246 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" event={"ID":"9d3537c0-52ed-4021-9d62-7f1027ae3f17","Type":"ContainerDied","Data":"a4a0e3b5f42b5e8cb8039f9c3f955adf9516ce30b5da1a93943ff018db50bc61"} Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.452262 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4a0e3b5f42b5e8cb8039f9c3f955adf9516ce30b5da1a93943ff018db50bc61" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.514941 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.568606 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-sb\") pod \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.568697 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-nb\") pod \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.568747 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gtr6\" (UniqueName: \"kubernetes.io/projected/9d3537c0-52ed-4021-9d62-7f1027ae3f17-kube-api-access-7gtr6\") pod \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.568811 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-config\") pod \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.568941 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-dns-svc\") pod \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\" (UID: \"9d3537c0-52ed-4021-9d62-7f1027ae3f17\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.583275 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3537c0-52ed-4021-9d62-7f1027ae3f17-kube-api-access-7gtr6" (OuterVolumeSpecName: "kube-api-access-7gtr6") pod "9d3537c0-52ed-4021-9d62-7f1027ae3f17" (UID: "9d3537c0-52ed-4021-9d62-7f1027ae3f17"). InnerVolumeSpecName "kube-api-access-7gtr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.621585 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d3537c0-52ed-4021-9d62-7f1027ae3f17" (UID: "9d3537c0-52ed-4021-9d62-7f1027ae3f17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.626103 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-config" (OuterVolumeSpecName: "config") pod "9d3537c0-52ed-4021-9d62-7f1027ae3f17" (UID: "9d3537c0-52ed-4021-9d62-7f1027ae3f17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.627499 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d3537c0-52ed-4021-9d62-7f1027ae3f17" (UID: "9d3537c0-52ed-4021-9d62-7f1027ae3f17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.649503 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d3537c0-52ed-4021-9d62-7f1027ae3f17" (UID: "9d3537c0-52ed-4021-9d62-7f1027ae3f17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.671473 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.671509 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.671521 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gtr6\" (UniqueName: \"kubernetes.io/projected/9d3537c0-52ed-4021-9d62-7f1027ae3f17-kube-api-access-7gtr6\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.671532 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.671540 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3537c0-52ed-4021-9d62-7f1027ae3f17-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.837662 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.843794 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.978536 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7h44\" (UniqueName: \"kubernetes.io/projected/06de8b69-fc4c-48b6-8fd6-6862a82b5620-kube-api-access-j7h44\") pod \"06de8b69-fc4c-48b6-8fd6-6862a82b5620\" (UID: \"06de8b69-fc4c-48b6-8fd6-6862a82b5620\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.978591 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq776\" (UniqueName: \"kubernetes.io/projected/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed-kube-api-access-jq776\") pod \"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed\" (UID: \"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed\") " Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.989245 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed-kube-api-access-jq776" (OuterVolumeSpecName: "kube-api-access-jq776") pod "cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed" (UID: "cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed"). InnerVolumeSpecName "kube-api-access-jq776". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:29 crc kubenswrapper[4796]: I0930 16:27:29.990717 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06de8b69-fc4c-48b6-8fd6-6862a82b5620-kube-api-access-j7h44" (OuterVolumeSpecName: "kube-api-access-j7h44") pod "06de8b69-fc4c-48b6-8fd6-6862a82b5620" (UID: "06de8b69-fc4c-48b6-8fd6-6862a82b5620"). InnerVolumeSpecName "kube-api-access-j7h44". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.080228 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7h44\" (UniqueName: \"kubernetes.io/projected/06de8b69-fc4c-48b6-8fd6-6862a82b5620-kube-api-access-j7h44\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.080265 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq776\" (UniqueName: \"kubernetes.io/projected/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed-kube-api-access-jq776\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.462283 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q2wwg" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.462269 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q2wwg" event={"ID":"cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed","Type":"ContainerDied","Data":"285b21ecb0dfdfc1aaa5a1f7c93a5d74a6b32530b9c583823f906018d6353cfc"} Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.462407 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="285b21ecb0dfdfc1aaa5a1f7c93a5d74a6b32530b9c583823f906018d6353cfc" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.463948 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z7lhm" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.464002 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l49w6" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.464027 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l49w6" event={"ID":"06de8b69-fc4c-48b6-8fd6-6862a82b5620","Type":"ContainerDied","Data":"bd2a40c98617bf675815e1c750d3f7e063146eba728e0434c76760f736adb0d8"} Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.464096 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd2a40c98617bf675815e1c750d3f7e063146eba728e0434c76760f736adb0d8" Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.516791 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z7lhm"] Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.530350 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z7lhm"] Sep 30 16:27:30 crc kubenswrapper[4796]: I0930 16:27:30.745551 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" path="/var/lib/kubelet/pods/9d3537c0-52ed-4021-9d62-7f1027ae3f17/volumes" Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.097798 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.097885 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.097954 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.098986 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f80108aae73ea16e4bbeae96cdafd6d40fe314e16173a204b191d54c90a06ba4"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.099127 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://f80108aae73ea16e4bbeae96cdafd6d40fe314e16173a204b191d54c90a06ba4" gracePeriod=600 Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.473878 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="f80108aae73ea16e4bbeae96cdafd6d40fe314e16173a204b191d54c90a06ba4" exitCode=0 Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.473927 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"f80108aae73ea16e4bbeae96cdafd6d40fe314e16173a204b191d54c90a06ba4"} Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.473961 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"f24d81f5faa8f6182a7d380f63e2c090bd17e98f5259e863fa184efdaf83296f"} Sep 30 16:27:31 crc kubenswrapper[4796]: I0930 16:27:31.474052 4796 scope.go:117] "RemoveContainer" containerID="8e292f0f61906ba72a5d11ef43cd6aabb03ee9b745e2fef2028fac6c22e38f4f" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.045554 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-dea9-account-create-t69m5"] Sep 30 16:27:32 crc kubenswrapper[4796]: E0930 16:27:32.046470 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerName="dnsmasq-dns" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046493 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerName="dnsmasq-dns" Sep 30 16:27:32 crc kubenswrapper[4796]: E0930 16:27:32.046507 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed" containerName="mariadb-database-create" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046519 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed" containerName="mariadb-database-create" Sep 30 16:27:32 crc kubenswrapper[4796]: E0930 16:27:32.046529 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06de8b69-fc4c-48b6-8fd6-6862a82b5620" containerName="mariadb-database-create" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046538 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="06de8b69-fc4c-48b6-8fd6-6862a82b5620" containerName="mariadb-database-create" Sep 30 16:27:32 crc kubenswrapper[4796]: E0930 16:27:32.046569 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerName="init" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046577 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerName="init" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046809 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed" containerName="mariadb-database-create" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046827 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="06de8b69-fc4c-48b6-8fd6-6862a82b5620" containerName="mariadb-database-create" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.046853 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3537c0-52ed-4021-9d62-7f1027ae3f17" containerName="dnsmasq-dns" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.047609 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.050511 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.057057 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-dea9-account-create-t69m5"] Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.125210 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htcg8\" (UniqueName: \"kubernetes.io/projected/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f-kube-api-access-htcg8\") pod \"glance-dea9-account-create-t69m5\" (UID: \"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f\") " pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.226809 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htcg8\" (UniqueName: \"kubernetes.io/projected/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f-kube-api-access-htcg8\") pod \"glance-dea9-account-create-t69m5\" (UID: \"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f\") " pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.248605 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htcg8\" (UniqueName: \"kubernetes.io/projected/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f-kube-api-access-htcg8\") pod \"glance-dea9-account-create-t69m5\" (UID: \"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f\") " pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.368831 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:32 crc kubenswrapper[4796]: W0930 16:27:32.856016 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4ec0db4_b42d_4e72_86ce_e8de9acbe63f.slice/crio-172cfef0b18e8a89205d48c1d1bd89fa80736649ae9cb19ef255c268557276c4 WatchSource:0}: Error finding container 172cfef0b18e8a89205d48c1d1bd89fa80736649ae9cb19ef255c268557276c4: Status 404 returned error can't find the container with id 172cfef0b18e8a89205d48c1d1bd89fa80736649ae9cb19ef255c268557276c4 Sep 30 16:27:32 crc kubenswrapper[4796]: I0930 16:27:32.863799 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-dea9-account-create-t69m5"] Sep 30 16:27:33 crc kubenswrapper[4796]: I0930 16:27:33.507781 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4ec0db4-b42d-4e72-86ce-e8de9acbe63f" containerID="b9217eed3cb3bbd9aafc49fc5bfe297476ae6026c8fbae38548b840469436ef7" exitCode=0 Sep 30 16:27:33 crc kubenswrapper[4796]: I0930 16:27:33.508541 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dea9-account-create-t69m5" event={"ID":"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f","Type":"ContainerDied","Data":"b9217eed3cb3bbd9aafc49fc5bfe297476ae6026c8fbae38548b840469436ef7"} Sep 30 16:27:33 crc kubenswrapper[4796]: I0930 16:27:33.508598 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dea9-account-create-t69m5" event={"ID":"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f","Type":"ContainerStarted","Data":"172cfef0b18e8a89205d48c1d1bd89fa80736649ae9cb19ef255c268557276c4"} Sep 30 16:27:34 crc kubenswrapper[4796]: I0930 16:27:34.522625 4796 generic.go:334] "Generic (PLEG): container finished" podID="420fc105-e180-4b89-af90-af130839e938" containerID="864b2c97e33185765c97cbaae5b9767ed575fcd7df71b1a7819a846d49f8236b" exitCode=0 Sep 30 16:27:34 crc kubenswrapper[4796]: I0930 16:27:34.522770 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6xqp" event={"ID":"420fc105-e180-4b89-af90-af130839e938","Type":"ContainerDied","Data":"864b2c97e33185765c97cbaae5b9767ed575fcd7df71b1a7819a846d49f8236b"} Sep 30 16:27:34 crc kubenswrapper[4796]: I0930 16:27:34.961453 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.074760 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htcg8\" (UniqueName: \"kubernetes.io/projected/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f-kube-api-access-htcg8\") pod \"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f\" (UID: \"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f\") " Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.080260 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f-kube-api-access-htcg8" (OuterVolumeSpecName: "kube-api-access-htcg8") pod "f4ec0db4-b42d-4e72-86ce-e8de9acbe63f" (UID: "f4ec0db4-b42d-4e72-86ce-e8de9acbe63f"). InnerVolumeSpecName "kube-api-access-htcg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.177316 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htcg8\" (UniqueName: \"kubernetes.io/projected/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f-kube-api-access-htcg8\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.539227 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dea9-account-create-t69m5" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.539299 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dea9-account-create-t69m5" event={"ID":"f4ec0db4-b42d-4e72-86ce-e8de9acbe63f","Type":"ContainerDied","Data":"172cfef0b18e8a89205d48c1d1bd89fa80736649ae9cb19ef255c268557276c4"} Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.539403 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="172cfef0b18e8a89205d48c1d1bd89fa80736649ae9cb19ef255c268557276c4" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.585971 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.595424 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/24528035-6f6d-4269-9e88-6171795db8a7-etc-swift\") pod \"swift-storage-0\" (UID: \"24528035-6f6d-4269-9e88-6171795db8a7\") " pod="openstack/swift-storage-0" Sep 30 16:27:35 crc kubenswrapper[4796]: I0930 16:27:35.604842 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.002809 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093527 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgz2j\" (UniqueName: \"kubernetes.io/projected/420fc105-e180-4b89-af90-af130839e938-kube-api-access-qgz2j\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093564 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-swiftconf\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093778 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-combined-ca-bundle\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093815 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-scripts\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093843 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-ring-data-devices\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093871 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-dispersionconf\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.093903 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/420fc105-e180-4b89-af90-af130839e938-etc-swift\") pod \"420fc105-e180-4b89-af90-af130839e938\" (UID: \"420fc105-e180-4b89-af90-af130839e938\") " Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.095285 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.095809 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/420fc105-e180-4b89-af90-af130839e938-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.105213 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420fc105-e180-4b89-af90-af130839e938-kube-api-access-qgz2j" (OuterVolumeSpecName: "kube-api-access-qgz2j") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "kube-api-access-qgz2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.107205 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.128163 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.133330 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.142279 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-scripts" (OuterVolumeSpecName: "scripts") pod "420fc105-e180-4b89-af90-af130839e938" (UID: "420fc105-e180-4b89-af90-af130839e938"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.183794 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195181 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195214 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195224 4796 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/420fc105-e180-4b89-af90-af130839e938-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195232 4796 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195241 4796 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/420fc105-e180-4b89-af90-af130839e938-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195250 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgz2j\" (UniqueName: \"kubernetes.io/projected/420fc105-e180-4b89-af90-af130839e938-kube-api-access-qgz2j\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.195263 4796 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/420fc105-e180-4b89-af90-af130839e938-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:36 crc kubenswrapper[4796]: W0930 16:27:36.258846 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24528035_6f6d_4269_9e88_6171795db8a7.slice/crio-feb45bd0cd99322b974c2497365099cd226ae5c96b4797e841196bd0f028aec1 WatchSource:0}: Error finding container feb45bd0cd99322b974c2497365099cd226ae5c96b4797e841196bd0f028aec1: Status 404 returned error can't find the container with id feb45bd0cd99322b974c2497365099cd226ae5c96b4797e841196bd0f028aec1 Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.494189 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8478-account-create-f9gfb"] Sep 30 16:27:36 crc kubenswrapper[4796]: E0930 16:27:36.494971 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ec0db4-b42d-4e72-86ce-e8de9acbe63f" containerName="mariadb-account-create" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.495047 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ec0db4-b42d-4e72-86ce-e8de9acbe63f" containerName="mariadb-account-create" Sep 30 16:27:36 crc kubenswrapper[4796]: E0930 16:27:36.495118 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420fc105-e180-4b89-af90-af130839e938" containerName="swift-ring-rebalance" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.495137 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="420fc105-e180-4b89-af90-af130839e938" containerName="swift-ring-rebalance" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.495549 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="420fc105-e180-4b89-af90-af130839e938" containerName="swift-ring-rebalance" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.495613 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ec0db4-b42d-4e72-86ce-e8de9acbe63f" containerName="mariadb-account-create" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.496869 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.500110 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.510498 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8478-account-create-f9gfb"] Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.549536 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"feb45bd0cd99322b974c2497365099cd226ae5c96b4797e841196bd0f028aec1"} Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.551908 4796 generic.go:334] "Generic (PLEG): container finished" podID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerID="8a60af1eaba786e386562dfc364d8f6a95c91b30228085291d68c87f3723546f" exitCode=0 Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.552036 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73","Type":"ContainerDied","Data":"8a60af1eaba786e386562dfc364d8f6a95c91b30228085291d68c87f3723546f"} Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.553952 4796 generic.go:334] "Generic (PLEG): container finished" podID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerID="677365ad12205b095e7ab1f45f8893d7344d64c584f08a009446e2d97c40ed89" exitCode=0 Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.554060 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee37a44-c6b6-4d75-ad41-9ff1402925f1","Type":"ContainerDied","Data":"677365ad12205b095e7ab1f45f8893d7344d64c584f08a009446e2d97c40ed89"} Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.557816 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6xqp" event={"ID":"420fc105-e180-4b89-af90-af130839e938","Type":"ContainerDied","Data":"630291bd6f4ef420f2f27a47428acdf3ed913463b6d691e5e277f1e842ee3c72"} Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.557852 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="630291bd6f4ef420f2f27a47428acdf3ed913463b6d691e5e277f1e842ee3c72" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.557903 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6xqp" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.605230 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7wkq\" (UniqueName: \"kubernetes.io/projected/59e52c40-093e-4b19-aa48-ce91889790aa-kube-api-access-b7wkq\") pod \"keystone-8478-account-create-f9gfb\" (UID: \"59e52c40-093e-4b19-aa48-ce91889790aa\") " pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.642655 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vfdx6" podUID="713f98aa-c3ec-4627-bc90-a1a577ca1c27" containerName="ovn-controller" probeResult="failure" output=< Sep 30 16:27:36 crc kubenswrapper[4796]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 16:27:36 crc kubenswrapper[4796]: > Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.706980 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7wkq\" (UniqueName: \"kubernetes.io/projected/59e52c40-093e-4b19-aa48-ce91889790aa-kube-api-access-b7wkq\") pod \"keystone-8478-account-create-f9gfb\" (UID: \"59e52c40-093e-4b19-aa48-ce91889790aa\") " pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.741226 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7wkq\" (UniqueName: \"kubernetes.io/projected/59e52c40-093e-4b19-aa48-ce91889790aa-kube-api-access-b7wkq\") pod \"keystone-8478-account-create-f9gfb\" (UID: \"59e52c40-093e-4b19-aa48-ce91889790aa\") " pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.776203 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-995c-account-create-bprn8"] Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.777472 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.779794 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.786761 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-995c-account-create-bprn8"] Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.809061 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc4cb\" (UniqueName: \"kubernetes.io/projected/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3-kube-api-access-zc4cb\") pod \"placement-995c-account-create-bprn8\" (UID: \"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3\") " pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.827796 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.916902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc4cb\" (UniqueName: \"kubernetes.io/projected/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3-kube-api-access-zc4cb\") pod \"placement-995c-account-create-bprn8\" (UID: \"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3\") " pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:36 crc kubenswrapper[4796]: I0930 16:27:36.937970 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc4cb\" (UniqueName: \"kubernetes.io/projected/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3-kube-api-access-zc4cb\") pod \"placement-995c-account-create-bprn8\" (UID: \"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3\") " pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.116207 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.211422 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-6bm8h"] Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.212585 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.219093 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.223284 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n2dwc" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.230218 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6bm8h"] Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.287537 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8478-account-create-f9gfb"] Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.322794 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-combined-ca-bundle\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.322913 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxtvv\" (UniqueName: \"kubernetes.io/projected/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-kube-api-access-zxtvv\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.323010 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-config-data\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.323086 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-db-sync-config-data\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.424876 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-config-data\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.424960 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-db-sync-config-data\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.425019 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-combined-ca-bundle\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.425060 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxtvv\" (UniqueName: \"kubernetes.io/projected/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-kube-api-access-zxtvv\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.430229 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-config-data\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.430235 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-combined-ca-bundle\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.430249 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-db-sync-config-data\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.439441 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxtvv\" (UniqueName: \"kubernetes.io/projected/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-kube-api-access-zxtvv\") pod \"glance-db-sync-6bm8h\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.544746 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bm8h" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.568848 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73","Type":"ContainerStarted","Data":"7fe663e8c846cefe85de508c57afcab325db9555b0e33b961273e60d34f8f45e"} Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.569107 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.574075 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee37a44-c6b6-4d75-ad41-9ff1402925f1","Type":"ContainerStarted","Data":"3f3ecd9dc8a6d7039567911cdd7c6d0fc7d9e82b6a0ea1a6e6a50ed719375835"} Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.574339 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.576173 4796 generic.go:334] "Generic (PLEG): container finished" podID="59e52c40-093e-4b19-aa48-ce91889790aa" containerID="016b0336b39e8f5b2712aec39f30b4e8ce36ae2a342735604b6b2804081f69cb" exitCode=0 Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.576220 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8478-account-create-f9gfb" event={"ID":"59e52c40-093e-4b19-aa48-ce91889790aa","Type":"ContainerDied","Data":"016b0336b39e8f5b2712aec39f30b4e8ce36ae2a342735604b6b2804081f69cb"} Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.576243 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8478-account-create-f9gfb" event={"ID":"59e52c40-093e-4b19-aa48-ce91889790aa","Type":"ContainerStarted","Data":"f3071973b14191bcd629d90d6335443bc6c5d3affcadd74f1ea0ff40ad483994"} Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.604998 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-995c-account-create-bprn8"] Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.607260 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=46.191889054 podStartE2EDuration="56.607245793s" podCreationTimestamp="2025-09-30 16:26:41 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.799054753 +0000 UTC m=+906.812333270" lastFinishedPulling="2025-09-30 16:27:05.214411482 +0000 UTC m=+917.227690009" observedRunningTime="2025-09-30 16:27:37.598681822 +0000 UTC m=+949.611960369" watchObservedRunningTime="2025-09-30 16:27:37.607245793 +0000 UTC m=+949.620524320" Sep 30 16:27:37 crc kubenswrapper[4796]: I0930 16:27:37.640826 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=47.525431305 podStartE2EDuration="55.640810246s" podCreationTimestamp="2025-09-30 16:26:42 +0000 UTC" firstStartedPulling="2025-09-30 16:26:54.784026902 +0000 UTC m=+906.797305439" lastFinishedPulling="2025-09-30 16:27:02.899405853 +0000 UTC m=+914.912684380" observedRunningTime="2025-09-30 16:27:37.637934232 +0000 UTC m=+949.651212769" watchObservedRunningTime="2025-09-30 16:27:37.640810246 +0000 UTC m=+949.654088773" Sep 30 16:27:37 crc kubenswrapper[4796]: W0930 16:27:37.787481 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb203ad29_5d4a_4cb6_9fca_ea89ec8576f3.slice/crio-0709e50b977ca20137b7ff517b7c6e4750aae064a02c20178c2691ef917ad835 WatchSource:0}: Error finding container 0709e50b977ca20137b7ff517b7c6e4750aae064a02c20178c2691ef917ad835: Status 404 returned error can't find the container with id 0709e50b977ca20137b7ff517b7c6e4750aae064a02c20178c2691ef917ad835 Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.362538 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6bm8h"] Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.588067 4796 generic.go:334] "Generic (PLEG): container finished" podID="b203ad29-5d4a-4cb6-9fca-ea89ec8576f3" containerID="0b6ca28cd7dab8dd7c3b3be9c24dab13c01715679360a50c2e672f5a2b0ceb05" exitCode=0 Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.588175 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-995c-account-create-bprn8" event={"ID":"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3","Type":"ContainerDied","Data":"0b6ca28cd7dab8dd7c3b3be9c24dab13c01715679360a50c2e672f5a2b0ceb05"} Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.588201 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-995c-account-create-bprn8" event={"ID":"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3","Type":"ContainerStarted","Data":"0709e50b977ca20137b7ff517b7c6e4750aae064a02c20178c2691ef917ad835"} Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.590869 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"920346a3dc474482486f742ff8c1360367852fa03197f89350f76fbb80324a9d"} Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.590939 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"308b849e5414a637d8ced5fa6001159b13849f591fd0f90624701b8c2796e210"} Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.590955 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"6b18cbfe36c3ab16884a46f6664acee87e15b194c1f265735a4ad4a93811612b"} Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.592917 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bm8h" event={"ID":"fe5beba9-e80d-45fd-90c8-72c7dbf2330f","Type":"ContainerStarted","Data":"ea2875883c1bcd1ab17cc5604002dfd840c512c93453d6dae35e171e4f0f65f6"} Sep 30 16:27:38 crc kubenswrapper[4796]: I0930 16:27:38.988406 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.055247 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7wkq\" (UniqueName: \"kubernetes.io/projected/59e52c40-093e-4b19-aa48-ce91889790aa-kube-api-access-b7wkq\") pod \"59e52c40-093e-4b19-aa48-ce91889790aa\" (UID: \"59e52c40-093e-4b19-aa48-ce91889790aa\") " Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.062108 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e52c40-093e-4b19-aa48-ce91889790aa-kube-api-access-b7wkq" (OuterVolumeSpecName: "kube-api-access-b7wkq") pod "59e52c40-093e-4b19-aa48-ce91889790aa" (UID: "59e52c40-093e-4b19-aa48-ce91889790aa"). InnerVolumeSpecName "kube-api-access-b7wkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.157378 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7wkq\" (UniqueName: \"kubernetes.io/projected/59e52c40-093e-4b19-aa48-ce91889790aa-kube-api-access-b7wkq\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.605653 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8478-account-create-f9gfb" event={"ID":"59e52c40-093e-4b19-aa48-ce91889790aa","Type":"ContainerDied","Data":"f3071973b14191bcd629d90d6335443bc6c5d3affcadd74f1ea0ff40ad483994"} Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.605717 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3071973b14191bcd629d90d6335443bc6c5d3affcadd74f1ea0ff40ad483994" Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.605743 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8478-account-create-f9gfb" Sep 30 16:27:39 crc kubenswrapper[4796]: I0930 16:27:39.609015 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"c1006d724c07ab61220edb10f2be837f55807e560f859ef8abf35be005be5dbc"} Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.094372 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.273873 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc4cb\" (UniqueName: \"kubernetes.io/projected/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3-kube-api-access-zc4cb\") pod \"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3\" (UID: \"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3\") " Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.281870 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3-kube-api-access-zc4cb" (OuterVolumeSpecName: "kube-api-access-zc4cb") pod "b203ad29-5d4a-4cb6-9fca-ea89ec8576f3" (UID: "b203ad29-5d4a-4cb6-9fca-ea89ec8576f3"). InnerVolumeSpecName "kube-api-access-zc4cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.376261 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc4cb\" (UniqueName: \"kubernetes.io/projected/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3-kube-api-access-zc4cb\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.619898 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"b69d2e993260606d7dc0fd542424401150ccacbc8b4b9b2ff70e0dec9c21ab5c"} Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.619934 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"20a0051387d98a96b218ce3dd985835a5d18af4570fe21fd0c18e0ff5df3658e"} Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.619944 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"e886982816d3f154aa7f2cbdc97a38c7c8b5381177467e437bca478736efacf6"} Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.622288 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-995c-account-create-bprn8" event={"ID":"b203ad29-5d4a-4cb6-9fca-ea89ec8576f3","Type":"ContainerDied","Data":"0709e50b977ca20137b7ff517b7c6e4750aae064a02c20178c2691ef917ad835"} Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.622310 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0709e50b977ca20137b7ff517b7c6e4750aae064a02c20178c2691ef917ad835" Sep 30 16:27:40 crc kubenswrapper[4796]: I0930 16:27:40.622352 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-995c-account-create-bprn8" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.640696 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"359e1e557b6fa17090eb53af76d22530fee623baa518d6e8df6fdf0535dbeb8f"} Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.641315 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vfdx6" podUID="713f98aa-c3ec-4627-bc90-a1a577ca1c27" containerName="ovn-controller" probeResult="failure" output=< Sep 30 16:27:41 crc kubenswrapper[4796]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 16:27:41 crc kubenswrapper[4796]: > Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.688122 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.698609 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lg6m7" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.906309 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vfdx6-config-ttlj2"] Sep 30 16:27:41 crc kubenswrapper[4796]: E0930 16:27:41.906829 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e52c40-093e-4b19-aa48-ce91889790aa" containerName="mariadb-account-create" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.906854 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e52c40-093e-4b19-aa48-ce91889790aa" containerName="mariadb-account-create" Sep 30 16:27:41 crc kubenswrapper[4796]: E0930 16:27:41.906870 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b203ad29-5d4a-4cb6-9fca-ea89ec8576f3" containerName="mariadb-account-create" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.906879 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b203ad29-5d4a-4cb6-9fca-ea89ec8576f3" containerName="mariadb-account-create" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.907140 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="59e52c40-093e-4b19-aa48-ce91889790aa" containerName="mariadb-account-create" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.907179 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b203ad29-5d4a-4cb6-9fca-ea89ec8576f3" containerName="mariadb-account-create" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.907894 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.910092 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 30 16:27:41 crc kubenswrapper[4796]: I0930 16:27:41.917506 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vfdx6-config-ttlj2"] Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.016348 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-additional-scripts\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.016437 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-scripts\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.016505 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run-ovn\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.016590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.016672 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-log-ovn\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.016697 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrqtr\" (UniqueName: \"kubernetes.io/projected/e1da9f22-6813-4058-b18c-28a6965de528-kube-api-access-vrqtr\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.117964 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-scripts\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118061 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run-ovn\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118112 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118172 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-log-ovn\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118193 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrqtr\" (UniqueName: \"kubernetes.io/projected/e1da9f22-6813-4058-b18c-28a6965de528-kube-api-access-vrqtr\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118215 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-additional-scripts\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118415 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run-ovn\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118419 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118425 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-log-ovn\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.118882 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-additional-scripts\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.120037 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-scripts\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.142677 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrqtr\" (UniqueName: \"kubernetes.io/projected/e1da9f22-6813-4058-b18c-28a6965de528-kube-api-access-vrqtr\") pod \"ovn-controller-vfdx6-config-ttlj2\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.234800 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.666819 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"976d525fddbda9edcef2f3ccbbb04dfb99371bc197fa01e3fe45aa168ac32a95"} Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.667245 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"de894711a113badaf1588846e625a2134d28fb074dd34a2db22abee8ce57cea1"} Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.667265 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"22cbf17fd6dcc5175cb304a635a22ce71002fb297fbd13212dbd39b83b450068"} Sep 30 16:27:42 crc kubenswrapper[4796]: I0930 16:27:42.757239 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vfdx6-config-ttlj2"] Sep 30 16:27:43 crc kubenswrapper[4796]: E0930 16:27:43.553586 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1da9f22_6813_4058_b18c_28a6965de528.slice/crio-fbcbd1746e285093161a27390765f3bb40791e8f88032b87aefdb854451cfff4.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:27:43 crc kubenswrapper[4796]: I0930 16:27:43.677070 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vfdx6-config-ttlj2" event={"ID":"e1da9f22-6813-4058-b18c-28a6965de528","Type":"ContainerStarted","Data":"fbcbd1746e285093161a27390765f3bb40791e8f88032b87aefdb854451cfff4"} Sep 30 16:27:43 crc kubenswrapper[4796]: I0930 16:27:43.679393 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vfdx6-config-ttlj2" event={"ID":"e1da9f22-6813-4058-b18c-28a6965de528","Type":"ContainerStarted","Data":"6d95ba9b165943f1c8318ff6e7288a27b04ccb040c6d3ff819779675b6ec430c"} Sep 30 16:27:43 crc kubenswrapper[4796]: I0930 16:27:43.683341 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"16f0e33321918b1c7441f7771a040106084f6a6a6f436c8c5b0126d77fc05d33"} Sep 30 16:27:43 crc kubenswrapper[4796]: I0930 16:27:43.683371 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"5ad1ea226ca46cc0b792a9905c286a4883c14eb99c3dc68f136b0f99cdcae33e"} Sep 30 16:27:43 crc kubenswrapper[4796]: I0930 16:27:43.713279 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vfdx6-config-ttlj2" podStartSLOduration=2.713252338 podStartE2EDuration="2.713252338s" podCreationTimestamp="2025-09-30 16:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:27:43.7037389 +0000 UTC m=+955.717017447" watchObservedRunningTime="2025-09-30 16:27:43.713252338 +0000 UTC m=+955.726530875" Sep 30 16:27:44 crc kubenswrapper[4796]: I0930 16:27:44.694334 4796 generic.go:334] "Generic (PLEG): container finished" podID="e1da9f22-6813-4058-b18c-28a6965de528" containerID="fbcbd1746e285093161a27390765f3bb40791e8f88032b87aefdb854451cfff4" exitCode=0 Sep 30 16:27:44 crc kubenswrapper[4796]: I0930 16:27:44.694413 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vfdx6-config-ttlj2" event={"ID":"e1da9f22-6813-4058-b18c-28a6965de528","Type":"ContainerDied","Data":"fbcbd1746e285093161a27390765f3bb40791e8f88032b87aefdb854451cfff4"} Sep 30 16:27:44 crc kubenswrapper[4796]: I0930 16:27:44.702582 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"c1d65ce143d2159c78e04d3b3e6dd8b701d109e1c643dd7826a5d9a5cb7f7c82"} Sep 30 16:27:46 crc kubenswrapper[4796]: I0930 16:27:46.631339 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vfdx6" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.537125 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590032 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-log-ovn\") pod \"e1da9f22-6813-4058-b18c-28a6965de528\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590085 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run-ovn\") pod \"e1da9f22-6813-4058-b18c-28a6965de528\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590121 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-additional-scripts\") pod \"e1da9f22-6813-4058-b18c-28a6965de528\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590131 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e1da9f22-6813-4058-b18c-28a6965de528" (UID: "e1da9f22-6813-4058-b18c-28a6965de528"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590171 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e1da9f22-6813-4058-b18c-28a6965de528" (UID: "e1da9f22-6813-4058-b18c-28a6965de528"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590187 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-scripts\") pod \"e1da9f22-6813-4058-b18c-28a6965de528\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590234 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrqtr\" (UniqueName: \"kubernetes.io/projected/e1da9f22-6813-4058-b18c-28a6965de528-kube-api-access-vrqtr\") pod \"e1da9f22-6813-4058-b18c-28a6965de528\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590259 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run\") pod \"e1da9f22-6813-4058-b18c-28a6965de528\" (UID: \"e1da9f22-6813-4058-b18c-28a6965de528\") " Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590738 4796 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590761 4796 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590787 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run" (OuterVolumeSpecName: "var-run") pod "e1da9f22-6813-4058-b18c-28a6965de528" (UID: "e1da9f22-6813-4058-b18c-28a6965de528"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.590798 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e1da9f22-6813-4058-b18c-28a6965de528" (UID: "e1da9f22-6813-4058-b18c-28a6965de528"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.591931 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-scripts" (OuterVolumeSpecName: "scripts") pod "e1da9f22-6813-4058-b18c-28a6965de528" (UID: "e1da9f22-6813-4058-b18c-28a6965de528"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.597235 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1da9f22-6813-4058-b18c-28a6965de528-kube-api-access-vrqtr" (OuterVolumeSpecName: "kube-api-access-vrqtr") pod "e1da9f22-6813-4058-b18c-28a6965de528" (UID: "e1da9f22-6813-4058-b18c-28a6965de528"). InnerVolumeSpecName "kube-api-access-vrqtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.692665 4796 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.692891 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1da9f22-6813-4058-b18c-28a6965de528-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.692901 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrqtr\" (UniqueName: \"kubernetes.io/projected/e1da9f22-6813-4058-b18c-28a6965de528-kube-api-access-vrqtr\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.692914 4796 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1da9f22-6813-4058-b18c-28a6965de528-var-run\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.767086 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vfdx6-config-ttlj2" event={"ID":"e1da9f22-6813-4058-b18c-28a6965de528","Type":"ContainerDied","Data":"6d95ba9b165943f1c8318ff6e7288a27b04ccb040c6d3ff819779675b6ec430c"} Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.767104 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vfdx6-config-ttlj2" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.767125 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d95ba9b165943f1c8318ff6e7288a27b04ccb040c6d3ff819779675b6ec430c" Sep 30 16:27:51 crc kubenswrapper[4796]: I0930 16:27:51.821635 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=28.02999457 podStartE2EDuration="33.821617994s" podCreationTimestamp="2025-09-30 16:27:18 +0000 UTC" firstStartedPulling="2025-09-30 16:27:36.261251626 +0000 UTC m=+948.274530153" lastFinishedPulling="2025-09-30 16:27:42.05287505 +0000 UTC m=+954.066153577" observedRunningTime="2025-09-30 16:27:51.814443594 +0000 UTC m=+963.827722151" watchObservedRunningTime="2025-09-30 16:27:51.821617994 +0000 UTC m=+963.834896521" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.067812 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-rw2xf"] Sep 30 16:27:52 crc kubenswrapper[4796]: E0930 16:27:52.068165 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1da9f22-6813-4058-b18c-28a6965de528" containerName="ovn-config" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.068181 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1da9f22-6813-4058-b18c-28a6965de528" containerName="ovn-config" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.068340 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1da9f22-6813-4058-b18c-28a6965de528" containerName="ovn-config" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.069708 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.071628 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.089211 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-rw2xf"] Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.200383 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.200716 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtcq8\" (UniqueName: \"kubernetes.io/projected/c25bb9b6-4c25-4011-982b-86d1622c7833-kube-api-access-qtcq8\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.200744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.200780 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-config\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.200814 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.201053 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.303339 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.303508 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.303669 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.303776 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtcq8\" (UniqueName: \"kubernetes.io/projected/c25bb9b6-4c25-4011-982b-86d1622c7833-kube-api-access-qtcq8\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.303821 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.303877 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-config\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.304813 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.304945 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.305154 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.305376 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-config\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.305538 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.322420 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtcq8\" (UniqueName: \"kubernetes.io/projected/c25bb9b6-4c25-4011-982b-86d1622c7833-kube-api-access-qtcq8\") pod \"dnsmasq-dns-77585f5f8c-rw2xf\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.385058 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.664316 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vfdx6-config-ttlj2"] Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.672327 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vfdx6-config-ttlj2"] Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.743877 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1da9f22-6813-4058-b18c-28a6965de528" path="/var/lib/kubelet/pods/e1da9f22-6813-4058-b18c-28a6965de528/volumes" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.805082 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"24528035-6f6d-4269-9e88-6171795db8a7","Type":"ContainerStarted","Data":"ee71827d4d55b1cffbc17e102f808392ff79278fc2e2e95fdc20aa4ee1286906"} Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.810059 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bm8h" event={"ID":"fe5beba9-e80d-45fd-90c8-72c7dbf2330f","Type":"ContainerStarted","Data":"1e1f238a04e2217ed3a436ae804da29e89f49e193166068df63f3d9af4b30cc0"} Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.833542 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-6bm8h" podStartSLOduration=2.619702488 podStartE2EDuration="15.833519373s" podCreationTimestamp="2025-09-30 16:27:37 +0000 UTC" firstStartedPulling="2025-09-30 16:27:38.377060488 +0000 UTC m=+950.390339015" lastFinishedPulling="2025-09-30 16:27:51.590877363 +0000 UTC m=+963.604155900" observedRunningTime="2025-09-30 16:27:52.82794813 +0000 UTC m=+964.841226677" watchObservedRunningTime="2025-09-30 16:27:52.833519373 +0000 UTC m=+964.846797910" Sep 30 16:27:52 crc kubenswrapper[4796]: I0930 16:27:52.869167 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-rw2xf"] Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.373157 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.654195 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.748598 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tfr8r"] Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.751408 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.757896 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tfr8r"] Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.826907 4796 generic.go:334] "Generic (PLEG): container finished" podID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerID="4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca" exitCode=0 Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.827057 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" event={"ID":"c25bb9b6-4c25-4011-982b-86d1622c7833","Type":"ContainerDied","Data":"4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca"} Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.827317 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" event={"ID":"c25bb9b6-4c25-4011-982b-86d1622c7833","Type":"ContainerStarted","Data":"5d69bb7d7be36b7d5634f56e983841a86789e09c4002fbb36fa8064e1ebcd7f7"} Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.903525 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-942bk"] Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.904920 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-942bk" Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.921132 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-942bk"] Sep 30 16:27:53 crc kubenswrapper[4796]: I0930 16:27:53.932571 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsk7s\" (UniqueName: \"kubernetes.io/projected/377e237f-78ad-4140-b988-690a2319c976-kube-api-access-gsk7s\") pod \"cinder-db-create-tfr8r\" (UID: \"377e237f-78ad-4140-b988-690a2319c976\") " pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.030639 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-6d4xh"] Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.032003 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.033856 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt8fw\" (UniqueName: \"kubernetes.io/projected/10c42976-648c-449e-9eb7-e876a4c3e67a-kube-api-access-pt8fw\") pod \"barbican-db-create-942bk\" (UID: \"10c42976-648c-449e-9eb7-e876a4c3e67a\") " pod="openstack/barbican-db-create-942bk" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.034100 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsk7s\" (UniqueName: \"kubernetes.io/projected/377e237f-78ad-4140-b988-690a2319c976-kube-api-access-gsk7s\") pod \"cinder-db-create-tfr8r\" (UID: \"377e237f-78ad-4140-b988-690a2319c976\") " pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.035317 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.035542 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.035936 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.036118 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxclq" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.049213 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6d4xh"] Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.055238 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsk7s\" (UniqueName: \"kubernetes.io/projected/377e237f-78ad-4140-b988-690a2319c976-kube-api-access-gsk7s\") pod \"cinder-db-create-tfr8r\" (UID: \"377e237f-78ad-4140-b988-690a2319c976\") " pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.092126 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.135838 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt8fw\" (UniqueName: \"kubernetes.io/projected/10c42976-648c-449e-9eb7-e876a4c3e67a-kube-api-access-pt8fw\") pod \"barbican-db-create-942bk\" (UID: \"10c42976-648c-449e-9eb7-e876a4c3e67a\") " pod="openstack/barbican-db-create-942bk" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.135896 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4xnl\" (UniqueName: \"kubernetes.io/projected/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-kube-api-access-f4xnl\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.135952 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-config-data\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.135988 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-combined-ca-bundle\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.153029 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9mss5"] Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.154338 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.157639 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt8fw\" (UniqueName: \"kubernetes.io/projected/10c42976-648c-449e-9eb7-e876a4c3e67a-kube-api-access-pt8fw\") pod \"barbican-db-create-942bk\" (UID: \"10c42976-648c-449e-9eb7-e876a4c3e67a\") " pod="openstack/barbican-db-create-942bk" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.169816 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9mss5"] Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.237028 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-config-data\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.237081 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-combined-ca-bundle\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.237188 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fsh6\" (UniqueName: \"kubernetes.io/projected/cc3dec6a-f05e-4320-a625-d89fdb79f2d4-kube-api-access-9fsh6\") pod \"neutron-db-create-9mss5\" (UID: \"cc3dec6a-f05e-4320-a625-d89fdb79f2d4\") " pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.237229 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4xnl\" (UniqueName: \"kubernetes.io/projected/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-kube-api-access-f4xnl\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.240890 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-config-data\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.243768 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-combined-ca-bundle\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.254355 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4xnl\" (UniqueName: \"kubernetes.io/projected/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-kube-api-access-f4xnl\") pod \"keystone-db-sync-6d4xh\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.304756 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-942bk" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.339055 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fsh6\" (UniqueName: \"kubernetes.io/projected/cc3dec6a-f05e-4320-a625-d89fdb79f2d4-kube-api-access-9fsh6\") pod \"neutron-db-create-9mss5\" (UID: \"cc3dec6a-f05e-4320-a625-d89fdb79f2d4\") " pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.374387 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.383714 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fsh6\" (UniqueName: \"kubernetes.io/projected/cc3dec6a-f05e-4320-a625-d89fdb79f2d4-kube-api-access-9fsh6\") pod \"neutron-db-create-9mss5\" (UID: \"cc3dec6a-f05e-4320-a625-d89fdb79f2d4\") " pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.472392 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.554290 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tfr8r"] Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.652397 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6d4xh"] Sep 30 16:27:54 crc kubenswrapper[4796]: W0930 16:27:54.664276 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4a5eee5_b362_4420_9cb1_29abf8a46d3a.slice/crio-8d6d9f64b3ae17c9bb23d2db202b621ae6711df0d37145690c21b13dd7bcf61f WatchSource:0}: Error finding container 8d6d9f64b3ae17c9bb23d2db202b621ae6711df0d37145690c21b13dd7bcf61f: Status 404 returned error can't find the container with id 8d6d9f64b3ae17c9bb23d2db202b621ae6711df0d37145690c21b13dd7bcf61f Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.836756 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" event={"ID":"c25bb9b6-4c25-4011-982b-86d1622c7833","Type":"ContainerStarted","Data":"90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954"} Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.836921 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.839692 4796 generic.go:334] "Generic (PLEG): container finished" podID="377e237f-78ad-4140-b988-690a2319c976" containerID="58bf3680142c54a61523f1fc71b893bc28f4670b5c4c2b2d246c1a717f1336d9" exitCode=0 Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.839846 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tfr8r" event={"ID":"377e237f-78ad-4140-b988-690a2319c976","Type":"ContainerDied","Data":"58bf3680142c54a61523f1fc71b893bc28f4670b5c4c2b2d246c1a717f1336d9"} Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.839871 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tfr8r" event={"ID":"377e237f-78ad-4140-b988-690a2319c976","Type":"ContainerStarted","Data":"30b5a3e99fbe2af8cc7a9c0aeae0095a8ea9752137e604eefa7a45eeab2922f1"} Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.841261 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6d4xh" event={"ID":"a4a5eee5-b362-4420-9cb1-29abf8a46d3a","Type":"ContainerStarted","Data":"8d6d9f64b3ae17c9bb23d2db202b621ae6711df0d37145690c21b13dd7bcf61f"} Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.858086 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" podStartSLOduration=2.8580671520000003 podStartE2EDuration="2.858067152s" podCreationTimestamp="2025-09-30 16:27:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:27:54.854189528 +0000 UTC m=+966.867468055" watchObservedRunningTime="2025-09-30 16:27:54.858067152 +0000 UTC m=+966.871345679" Sep 30 16:27:54 crc kubenswrapper[4796]: W0930 16:27:54.936628 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c42976_648c_449e_9eb7_e876a4c3e67a.slice/crio-d5eef95bd62437ec224225e1b811c974fd3c7d7b2cd17ca9f1b128b2d0a8546c WatchSource:0}: Error finding container d5eef95bd62437ec224225e1b811c974fd3c7d7b2cd17ca9f1b128b2d0a8546c: Status 404 returned error can't find the container with id d5eef95bd62437ec224225e1b811c974fd3c7d7b2cd17ca9f1b128b2d0a8546c Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.936871 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-942bk"] Sep 30 16:27:54 crc kubenswrapper[4796]: I0930 16:27:54.983606 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9mss5"] Sep 30 16:27:54 crc kubenswrapper[4796]: W0930 16:27:54.991002 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc3dec6a_f05e_4320_a625_d89fdb79f2d4.slice/crio-df85277599d93eaaad2cfbe8b6f853a11307e57a36f9decdb48381e2bff8dae5 WatchSource:0}: Error finding container df85277599d93eaaad2cfbe8b6f853a11307e57a36f9decdb48381e2bff8dae5: Status 404 returned error can't find the container with id df85277599d93eaaad2cfbe8b6f853a11307e57a36f9decdb48381e2bff8dae5 Sep 30 16:27:55 crc kubenswrapper[4796]: I0930 16:27:55.850475 4796 generic.go:334] "Generic (PLEG): container finished" podID="10c42976-648c-449e-9eb7-e876a4c3e67a" containerID="feb7ef30d37b7763c79a804fab949915cc8a68ea9292fc9399bc71c8e95620e5" exitCode=0 Sep 30 16:27:55 crc kubenswrapper[4796]: I0930 16:27:55.850764 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-942bk" event={"ID":"10c42976-648c-449e-9eb7-e876a4c3e67a","Type":"ContainerDied","Data":"feb7ef30d37b7763c79a804fab949915cc8a68ea9292fc9399bc71c8e95620e5"} Sep 30 16:27:55 crc kubenswrapper[4796]: I0930 16:27:55.850798 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-942bk" event={"ID":"10c42976-648c-449e-9eb7-e876a4c3e67a","Type":"ContainerStarted","Data":"d5eef95bd62437ec224225e1b811c974fd3c7d7b2cd17ca9f1b128b2d0a8546c"} Sep 30 16:27:55 crc kubenswrapper[4796]: I0930 16:27:55.852411 4796 generic.go:334] "Generic (PLEG): container finished" podID="cc3dec6a-f05e-4320-a625-d89fdb79f2d4" containerID="98b3b3af45327799ac7ad8d353f6163d748490a9e8f7e6181a055df49f6a6b7f" exitCode=0 Sep 30 16:27:55 crc kubenswrapper[4796]: I0930 16:27:55.852512 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9mss5" event={"ID":"cc3dec6a-f05e-4320-a625-d89fdb79f2d4","Type":"ContainerDied","Data":"98b3b3af45327799ac7ad8d353f6163d748490a9e8f7e6181a055df49f6a6b7f"} Sep 30 16:27:55 crc kubenswrapper[4796]: I0930 16:27:55.852567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9mss5" event={"ID":"cc3dec6a-f05e-4320-a625-d89fdb79f2d4","Type":"ContainerStarted","Data":"df85277599d93eaaad2cfbe8b6f853a11307e57a36f9decdb48381e2bff8dae5"} Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.193510 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.278367 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsk7s\" (UniqueName: \"kubernetes.io/projected/377e237f-78ad-4140-b988-690a2319c976-kube-api-access-gsk7s\") pod \"377e237f-78ad-4140-b988-690a2319c976\" (UID: \"377e237f-78ad-4140-b988-690a2319c976\") " Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.283550 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377e237f-78ad-4140-b988-690a2319c976-kube-api-access-gsk7s" (OuterVolumeSpecName: "kube-api-access-gsk7s") pod "377e237f-78ad-4140-b988-690a2319c976" (UID: "377e237f-78ad-4140-b988-690a2319c976"). InnerVolumeSpecName "kube-api-access-gsk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.381034 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsk7s\" (UniqueName: \"kubernetes.io/projected/377e237f-78ad-4140-b988-690a2319c976-kube-api-access-gsk7s\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.867370 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tfr8r" Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.867420 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tfr8r" event={"ID":"377e237f-78ad-4140-b988-690a2319c976","Type":"ContainerDied","Data":"30b5a3e99fbe2af8cc7a9c0aeae0095a8ea9752137e604eefa7a45eeab2922f1"} Sep 30 16:27:56 crc kubenswrapper[4796]: I0930 16:27:56.867447 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30b5a3e99fbe2af8cc7a9c0aeae0095a8ea9752137e604eefa7a45eeab2922f1" Sep 30 16:27:58 crc kubenswrapper[4796]: I0930 16:27:58.884835 4796 generic.go:334] "Generic (PLEG): container finished" podID="fe5beba9-e80d-45fd-90c8-72c7dbf2330f" containerID="1e1f238a04e2217ed3a436ae804da29e89f49e193166068df63f3d9af4b30cc0" exitCode=0 Sep 30 16:27:58 crc kubenswrapper[4796]: I0930 16:27:58.884889 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bm8h" event={"ID":"fe5beba9-e80d-45fd-90c8-72c7dbf2330f","Type":"ContainerDied","Data":"1e1f238a04e2217ed3a436ae804da29e89f49e193166068df63f3d9af4b30cc0"} Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.207113 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-942bk" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.213599 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.329527 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fsh6\" (UniqueName: \"kubernetes.io/projected/cc3dec6a-f05e-4320-a625-d89fdb79f2d4-kube-api-access-9fsh6\") pod \"cc3dec6a-f05e-4320-a625-d89fdb79f2d4\" (UID: \"cc3dec6a-f05e-4320-a625-d89fdb79f2d4\") " Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.329579 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt8fw\" (UniqueName: \"kubernetes.io/projected/10c42976-648c-449e-9eb7-e876a4c3e67a-kube-api-access-pt8fw\") pod \"10c42976-648c-449e-9eb7-e876a4c3e67a\" (UID: \"10c42976-648c-449e-9eb7-e876a4c3e67a\") " Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.333329 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3dec6a-f05e-4320-a625-d89fdb79f2d4-kube-api-access-9fsh6" (OuterVolumeSpecName: "kube-api-access-9fsh6") pod "cc3dec6a-f05e-4320-a625-d89fdb79f2d4" (UID: "cc3dec6a-f05e-4320-a625-d89fdb79f2d4"). InnerVolumeSpecName "kube-api-access-9fsh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.333423 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c42976-648c-449e-9eb7-e876a4c3e67a-kube-api-access-pt8fw" (OuterVolumeSpecName: "kube-api-access-pt8fw") pod "10c42976-648c-449e-9eb7-e876a4c3e67a" (UID: "10c42976-648c-449e-9eb7-e876a4c3e67a"). InnerVolumeSpecName "kube-api-access-pt8fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.430784 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fsh6\" (UniqueName: \"kubernetes.io/projected/cc3dec6a-f05e-4320-a625-d89fdb79f2d4-kube-api-access-9fsh6\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.430812 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt8fw\" (UniqueName: \"kubernetes.io/projected/10c42976-648c-449e-9eb7-e876a4c3e67a-kube-api-access-pt8fw\") on node \"crc\" DevicePath \"\"" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.900198 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-942bk" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.900199 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-942bk" event={"ID":"10c42976-648c-449e-9eb7-e876a4c3e67a","Type":"ContainerDied","Data":"d5eef95bd62437ec224225e1b811c974fd3c7d7b2cd17ca9f1b128b2d0a8546c"} Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.900962 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5eef95bd62437ec224225e1b811c974fd3c7d7b2cd17ca9f1b128b2d0a8546c" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.903431 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9mss5" event={"ID":"cc3dec6a-f05e-4320-a625-d89fdb79f2d4","Type":"ContainerDied","Data":"df85277599d93eaaad2cfbe8b6f853a11307e57a36f9decdb48381e2bff8dae5"} Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.903499 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df85277599d93eaaad2cfbe8b6f853a11307e57a36f9decdb48381e2bff8dae5" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.903680 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9mss5" Sep 30 16:27:59 crc kubenswrapper[4796]: I0930 16:27:59.906472 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6d4xh" event={"ID":"a4a5eee5-b362-4420-9cb1-29abf8a46d3a","Type":"ContainerStarted","Data":"4a9955c3325e12f6a19b92fbb7386c490103157c5c5200a90824b06993dd16a4"} Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.268882 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-6d4xh" podStartSLOduration=1.885802526 podStartE2EDuration="6.268859119s" podCreationTimestamp="2025-09-30 16:27:54 +0000 UTC" firstStartedPulling="2025-09-30 16:27:54.671038472 +0000 UTC m=+966.684316999" lastFinishedPulling="2025-09-30 16:27:59.054095055 +0000 UTC m=+971.067373592" observedRunningTime="2025-09-30 16:27:59.940462576 +0000 UTC m=+971.953741153" watchObservedRunningTime="2025-09-30 16:28:00.268859119 +0000 UTC m=+972.282137656" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.424178 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bm8h" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.553395 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-config-data\") pod \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.553497 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-combined-ca-bundle\") pod \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.553639 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-db-sync-config-data\") pod \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.553722 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxtvv\" (UniqueName: \"kubernetes.io/projected/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-kube-api-access-zxtvv\") pod \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\" (UID: \"fe5beba9-e80d-45fd-90c8-72c7dbf2330f\") " Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.560520 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-kube-api-access-zxtvv" (OuterVolumeSpecName: "kube-api-access-zxtvv") pod "fe5beba9-e80d-45fd-90c8-72c7dbf2330f" (UID: "fe5beba9-e80d-45fd-90c8-72c7dbf2330f"). InnerVolumeSpecName "kube-api-access-zxtvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.561471 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fe5beba9-e80d-45fd-90c8-72c7dbf2330f" (UID: "fe5beba9-e80d-45fd-90c8-72c7dbf2330f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.587527 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe5beba9-e80d-45fd-90c8-72c7dbf2330f" (UID: "fe5beba9-e80d-45fd-90c8-72c7dbf2330f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.630540 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-config-data" (OuterVolumeSpecName: "config-data") pod "fe5beba9-e80d-45fd-90c8-72c7dbf2330f" (UID: "fe5beba9-e80d-45fd-90c8-72c7dbf2330f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.657006 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxtvv\" (UniqueName: \"kubernetes.io/projected/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-kube-api-access-zxtvv\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.657046 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.657060 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.657072 4796 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe5beba9-e80d-45fd-90c8-72c7dbf2330f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.920148 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bm8h" event={"ID":"fe5beba9-e80d-45fd-90c8-72c7dbf2330f","Type":"ContainerDied","Data":"ea2875883c1bcd1ab17cc5604002dfd840c512c93453d6dae35e171e4f0f65f6"} Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.920236 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea2875883c1bcd1ab17cc5604002dfd840c512c93453d6dae35e171e4f0f65f6" Sep 30 16:28:00 crc kubenswrapper[4796]: I0930 16:28:00.920420 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bm8h" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.218617 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-rw2xf"] Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.221031 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.224768 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerName="dnsmasq-dns" containerID="cri-o://90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954" gracePeriod=10 Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.273275 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-l6lb8"] Sep 30 16:28:01 crc kubenswrapper[4796]: E0930 16:28:01.273712 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe5beba9-e80d-45fd-90c8-72c7dbf2330f" containerName="glance-db-sync" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.273746 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe5beba9-e80d-45fd-90c8-72c7dbf2330f" containerName="glance-db-sync" Sep 30 16:28:01 crc kubenswrapper[4796]: E0930 16:28:01.273775 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377e237f-78ad-4140-b988-690a2319c976" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.273784 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="377e237f-78ad-4140-b988-690a2319c976" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: E0930 16:28:01.273810 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c42976-648c-449e-9eb7-e876a4c3e67a" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.273818 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c42976-648c-449e-9eb7-e876a4c3e67a" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: E0930 16:28:01.273833 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3dec6a-f05e-4320-a625-d89fdb79f2d4" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.273840 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3dec6a-f05e-4320-a625-d89fdb79f2d4" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.274082 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3dec6a-f05e-4320-a625-d89fdb79f2d4" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.274106 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe5beba9-e80d-45fd-90c8-72c7dbf2330f" containerName="glance-db-sync" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.274123 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c42976-648c-449e-9eb7-e876a4c3e67a" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.274140 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="377e237f-78ad-4140-b988-690a2319c976" containerName="mariadb-database-create" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.275302 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.287016 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-l6lb8"] Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.369050 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.369382 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.369410 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.369432 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr6t6\" (UniqueName: \"kubernetes.io/projected/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-kube-api-access-rr6t6\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.369486 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-config\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.369511 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.470964 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-config\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.471054 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.471122 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.471145 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.471169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.471189 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr6t6\" (UniqueName: \"kubernetes.io/projected/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-kube-api-access-rr6t6\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.472238 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-config\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.472748 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.473117 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.473448 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.475765 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.489595 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr6t6\" (UniqueName: \"kubernetes.io/projected/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-kube-api-access-rr6t6\") pod \"dnsmasq-dns-7ff5475cc9-l6lb8\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.679742 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.698087 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.774800 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-config\") pod \"c25bb9b6-4c25-4011-982b-86d1622c7833\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.775207 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-sb\") pod \"c25bb9b6-4c25-4011-982b-86d1622c7833\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.775233 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-swift-storage-0\") pod \"c25bb9b6-4c25-4011-982b-86d1622c7833\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.775267 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-nb\") pod \"c25bb9b6-4c25-4011-982b-86d1622c7833\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.775296 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-svc\") pod \"c25bb9b6-4c25-4011-982b-86d1622c7833\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.775366 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtcq8\" (UniqueName: \"kubernetes.io/projected/c25bb9b6-4c25-4011-982b-86d1622c7833-kube-api-access-qtcq8\") pod \"c25bb9b6-4c25-4011-982b-86d1622c7833\" (UID: \"c25bb9b6-4c25-4011-982b-86d1622c7833\") " Sep 30 16:28:01 crc kubenswrapper[4796]: I0930 16:28:01.781681 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c25bb9b6-4c25-4011-982b-86d1622c7833-kube-api-access-qtcq8" (OuterVolumeSpecName: "kube-api-access-qtcq8") pod "c25bb9b6-4c25-4011-982b-86d1622c7833" (UID: "c25bb9b6-4c25-4011-982b-86d1622c7833"). InnerVolumeSpecName "kube-api-access-qtcq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.829593 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-config" (OuterVolumeSpecName: "config") pod "c25bb9b6-4c25-4011-982b-86d1622c7833" (UID: "c25bb9b6-4c25-4011-982b-86d1622c7833"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.832385 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c25bb9b6-4c25-4011-982b-86d1622c7833" (UID: "c25bb9b6-4c25-4011-982b-86d1622c7833"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.835631 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c25bb9b6-4c25-4011-982b-86d1622c7833" (UID: "c25bb9b6-4c25-4011-982b-86d1622c7833"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.835738 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c25bb9b6-4c25-4011-982b-86d1622c7833" (UID: "c25bb9b6-4c25-4011-982b-86d1622c7833"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.865665 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c25bb9b6-4c25-4011-982b-86d1622c7833" (UID: "c25bb9b6-4c25-4011-982b-86d1622c7833"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.877280 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.877307 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.877315 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.877324 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.877334 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtcq8\" (UniqueName: \"kubernetes.io/projected/c25bb9b6-4c25-4011-982b-86d1622c7833-kube-api-access-qtcq8\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.877343 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c25bb9b6-4c25-4011-982b-86d1622c7833-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.941930 4796 generic.go:334] "Generic (PLEG): container finished" podID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerID="90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954" exitCode=0 Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.941972 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" event={"ID":"c25bb9b6-4c25-4011-982b-86d1622c7833","Type":"ContainerDied","Data":"90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954"} Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.942023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" event={"ID":"c25bb9b6-4c25-4011-982b-86d1622c7833","Type":"ContainerDied","Data":"5d69bb7d7be36b7d5634f56e983841a86789e09c4002fbb36fa8064e1ebcd7f7"} Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.942040 4796 scope.go:117] "RemoveContainer" containerID="90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.942129 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-rw2xf" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.969777 4796 scope.go:117] "RemoveContainer" containerID="4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.988046 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-rw2xf"] Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:01.993765 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-rw2xf"] Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.009709 4796 scope.go:117] "RemoveContainer" containerID="90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954" Sep 30 16:28:02 crc kubenswrapper[4796]: E0930 16:28:02.010203 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954\": container with ID starting with 90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954 not found: ID does not exist" containerID="90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.010252 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954"} err="failed to get container status \"90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954\": rpc error: code = NotFound desc = could not find container \"90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954\": container with ID starting with 90e6b48e71960a86a15aff3deccfed26743ab7c7645ad537754922e3a4cc3954 not found: ID does not exist" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.010278 4796 scope.go:117] "RemoveContainer" containerID="4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca" Sep 30 16:28:02 crc kubenswrapper[4796]: E0930 16:28:02.010590 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca\": container with ID starting with 4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca not found: ID does not exist" containerID="4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.010629 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca"} err="failed to get container status \"4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca\": rpc error: code = NotFound desc = could not find container \"4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca\": container with ID starting with 4e557ad19efa7155c23f22c656a6e0dcfabe83c65ecbecc7f0145782023a30ca not found: ID does not exist" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.744911 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" path="/var/lib/kubelet/pods/c25bb9b6-4c25-4011-982b-86d1622c7833/volumes" Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.765860 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-l6lb8"] Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.957175 4796 generic.go:334] "Generic (PLEG): container finished" podID="a4a5eee5-b362-4420-9cb1-29abf8a46d3a" containerID="4a9955c3325e12f6a19b92fbb7386c490103157c5c5200a90824b06993dd16a4" exitCode=0 Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.957258 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6d4xh" event={"ID":"a4a5eee5-b362-4420-9cb1-29abf8a46d3a","Type":"ContainerDied","Data":"4a9955c3325e12f6a19b92fbb7386c490103157c5c5200a90824b06993dd16a4"} Sep 30 16:28:02 crc kubenswrapper[4796]: I0930 16:28:02.958802 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" event={"ID":"cd0fbb84-5d18-47bf-a0e9-0193eb580b01","Type":"ContainerStarted","Data":"5fadeae05b52c6e1ac95224c99281f2bd0488351ec67d2761e033a1e5209afee"} Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.784604 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-96a6-account-create-xfrx4"] Sep 30 16:28:03 crc kubenswrapper[4796]: E0930 16:28:03.785113 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerName="init" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.785128 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerName="init" Sep 30 16:28:03 crc kubenswrapper[4796]: E0930 16:28:03.785139 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerName="dnsmasq-dns" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.785145 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerName="dnsmasq-dns" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.785349 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c25bb9b6-4c25-4011-982b-86d1622c7833" containerName="dnsmasq-dns" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.785915 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.787641 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.800527 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-96a6-account-create-xfrx4"] Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.929330 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8bv6\" (UniqueName: \"kubernetes.io/projected/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f-kube-api-access-l8bv6\") pod \"cinder-96a6-account-create-xfrx4\" (UID: \"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f\") " pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.972022 4796 generic.go:334] "Generic (PLEG): container finished" podID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerID="4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023" exitCode=0 Sep 30 16:28:03 crc kubenswrapper[4796]: I0930 16:28:03.972105 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" event={"ID":"cd0fbb84-5d18-47bf-a0e9-0193eb580b01","Type":"ContainerDied","Data":"4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023"} Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.030743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8bv6\" (UniqueName: \"kubernetes.io/projected/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f-kube-api-access-l8bv6\") pod \"cinder-96a6-account-create-xfrx4\" (UID: \"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f\") " pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.048586 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8bv6\" (UniqueName: \"kubernetes.io/projected/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f-kube-api-access-l8bv6\") pod \"cinder-96a6-account-create-xfrx4\" (UID: \"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f\") " pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.148447 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.300603 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.436549 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4xnl\" (UniqueName: \"kubernetes.io/projected/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-kube-api-access-f4xnl\") pod \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.436860 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-combined-ca-bundle\") pod \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.437679 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-config-data\") pod \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\" (UID: \"a4a5eee5-b362-4420-9cb1-29abf8a46d3a\") " Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.448720 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-kube-api-access-f4xnl" (OuterVolumeSpecName: "kube-api-access-f4xnl") pod "a4a5eee5-b362-4420-9cb1-29abf8a46d3a" (UID: "a4a5eee5-b362-4420-9cb1-29abf8a46d3a"). InnerVolumeSpecName "kube-api-access-f4xnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.473826 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4a5eee5-b362-4420-9cb1-29abf8a46d3a" (UID: "a4a5eee5-b362-4420-9cb1-29abf8a46d3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.485518 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-96a6-account-create-xfrx4"] Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.486257 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-config-data" (OuterVolumeSpecName: "config-data") pod "a4a5eee5-b362-4420-9cb1-29abf8a46d3a" (UID: "a4a5eee5-b362-4420-9cb1-29abf8a46d3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.539385 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4xnl\" (UniqueName: \"kubernetes.io/projected/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-kube-api-access-f4xnl\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.539430 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.539442 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a5eee5-b362-4420-9cb1-29abf8a46d3a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.985843 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6d4xh" event={"ID":"a4a5eee5-b362-4420-9cb1-29abf8a46d3a","Type":"ContainerDied","Data":"8d6d9f64b3ae17c9bb23d2db202b621ae6711df0d37145690c21b13dd7bcf61f"} Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.985894 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d6d9f64b3ae17c9bb23d2db202b621ae6711df0d37145690c21b13dd7bcf61f" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.986004 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6d4xh" Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.988020 4796 generic.go:334] "Generic (PLEG): container finished" podID="b39aff1c-79f0-4747-a111-7c5ac3a7dc3f" containerID="884af2dce8fbb96f9fc0a188df28b12eec1735e11851391a7b8c5f03fff8e547" exitCode=0 Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.988159 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-96a6-account-create-xfrx4" event={"ID":"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f","Type":"ContainerDied","Data":"884af2dce8fbb96f9fc0a188df28b12eec1735e11851391a7b8c5f03fff8e547"} Sep 30 16:28:04 crc kubenswrapper[4796]: I0930 16:28:04.988252 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-96a6-account-create-xfrx4" event={"ID":"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f","Type":"ContainerStarted","Data":"3d9e46ba669c4982462bdfaad6c5106e1af34bdba88725fb98d1a61010991b4b"} Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:04.998718 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" event={"ID":"cd0fbb84-5d18-47bf-a0e9-0193eb580b01","Type":"ContainerStarted","Data":"e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc"} Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:04.999557 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.039554 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" podStartSLOduration=4.039535319 podStartE2EDuration="4.039535319s" podCreationTimestamp="2025-09-30 16:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:05.028923608 +0000 UTC m=+977.042202155" watchObservedRunningTime="2025-09-30 16:28:05.039535319 +0000 UTC m=+977.052813846" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.229329 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-l6lb8"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.266905 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7bmxm"] Sep 30 16:28:05 crc kubenswrapper[4796]: E0930 16:28:05.267430 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a5eee5-b362-4420-9cb1-29abf8a46d3a" containerName="keystone-db-sync" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.267497 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a5eee5-b362-4420-9cb1-29abf8a46d3a" containerName="keystone-db-sync" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.272090 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a5eee5-b362-4420-9cb1-29abf8a46d3a" containerName="keystone-db-sync" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.272906 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.276273 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.277968 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.280337 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.280395 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.280481 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.283334 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxclq" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.299288 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7bmxm"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.312134 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355218 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nqrr\" (UniqueName: \"kubernetes.io/projected/3d32d14f-94c5-4714-823a-8875a3d840db-kube-api-access-5nqrr\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-config-data\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355312 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-combined-ca-bundle\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355340 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355378 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355398 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptlps\" (UniqueName: \"kubernetes.io/projected/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-kube-api-access-ptlps\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355424 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355448 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-config\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355466 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-fernet-keys\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355505 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-credential-keys\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355573 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.355602 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-scripts\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457234 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptlps\" (UniqueName: \"kubernetes.io/projected/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-kube-api-access-ptlps\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457292 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457321 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-config\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457343 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-fernet-keys\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457365 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-credential-keys\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457417 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457441 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-scripts\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457510 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nqrr\" (UniqueName: \"kubernetes.io/projected/3d32d14f-94c5-4714-823a-8875a3d840db-kube-api-access-5nqrr\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457551 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-config-data\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457556 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457573 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-combined-ca-bundle\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457597 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.457639 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.458503 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.458546 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.459096 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.459232 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.459736 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-config\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.460767 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.464015 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-combined-ca-bundle\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.464732 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-config-data\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.465352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-credential-keys\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.466265 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.466610 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.469604 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-fernet-keys\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.478590 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.482394 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-scripts\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.482731 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nqrr\" (UniqueName: \"kubernetes.io/projected/3d32d14f-94c5-4714-823a-8875a3d840db-kube-api-access-5nqrr\") pod \"keystone-bootstrap-7bmxm\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.489112 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptlps\" (UniqueName: \"kubernetes.io/projected/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-kube-api-access-ptlps\") pod \"dnsmasq-dns-5c5cc7c5ff-tzz7t\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559232 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-log-httpd\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559279 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdnjh\" (UniqueName: \"kubernetes.io/projected/8c4f9d73-18f6-4133-87c7-e120155cf81c-kube-api-access-qdnjh\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-run-httpd\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559353 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559409 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-scripts\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559425 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-config-data\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.559458 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.608323 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.632299 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.656391 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660643 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-scripts\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660684 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-config-data\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660724 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660754 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-log-httpd\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660780 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdnjh\" (UniqueName: \"kubernetes.io/projected/8c4f9d73-18f6-4133-87c7-e120155cf81c-kube-api-access-qdnjh\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660820 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-run-httpd\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.660845 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.665794 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-log-httpd\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.666024 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-run-httpd\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.669999 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.683091 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-config-data\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.683991 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xp6sc"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.686195 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.686750 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.688274 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bv65w" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.688427 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.691625 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-scripts\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.694111 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdnjh\" (UniqueName: \"kubernetes.io/projected/8c4f9d73-18f6-4133-87c7-e120155cf81c-kube-api-access-qdnjh\") pod \"ceilometer-0\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.694561 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.705026 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xp6sc"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.713636 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gclst"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.715475 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.738176 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gclst"] Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.764116 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-config-data\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.764156 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlvgf\" (UniqueName: \"kubernetes.io/projected/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-kube-api-access-mlvgf\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.764240 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-combined-ca-bundle\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.764269 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-scripts\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.764295 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-logs\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867142 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-config-data\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867177 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlvgf\" (UniqueName: \"kubernetes.io/projected/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-kube-api-access-mlvgf\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867261 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867296 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867324 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-config\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867346 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-combined-ca-bundle\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867369 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-scripts\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867392 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxgdq\" (UniqueName: \"kubernetes.io/projected/165b3d32-fe2a-47a1-8adf-a70222767c2c-kube-api-access-nxgdq\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867412 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-logs\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.867429 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.868518 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-logs\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.872531 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-config-data\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.872746 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-combined-ca-bundle\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.875836 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-scripts\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.880850 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.883988 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlvgf\" (UniqueName: \"kubernetes.io/projected/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-kube-api-access-mlvgf\") pod \"placement-db-sync-xp6sc\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.969069 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-config\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.969362 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxgdq\" (UniqueName: \"kubernetes.io/projected/165b3d32-fe2a-47a1-8adf-a70222767c2c-kube-api-access-nxgdq\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.969387 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.969456 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.969531 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.969553 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.970249 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.970308 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-config\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.970525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.970729 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.970874 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:05 crc kubenswrapper[4796]: I0930 16:28:05.994563 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxgdq\" (UniqueName: \"kubernetes.io/projected/165b3d32-fe2a-47a1-8adf-a70222767c2c-kube-api-access-nxgdq\") pod \"dnsmasq-dns-8b5c85b87-gclst\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.080345 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.087965 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.192153 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.200067 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7bmxm"] Sep 30 16:28:06 crc kubenswrapper[4796]: W0930 16:28:06.240420 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b954a1b_4ce8_46ac_bb4e_cac178f07f25.slice/crio-a0c7e36772890a83f4699a02b4db631c8eb014fb72b9a3040791c21f3c6b5cfe WatchSource:0}: Error finding container a0c7e36772890a83f4699a02b4db631c8eb014fb72b9a3040791c21f3c6b5cfe: Status 404 returned error can't find the container with id a0c7e36772890a83f4699a02b4db631c8eb014fb72b9a3040791c21f3c6b5cfe Sep 30 16:28:06 crc kubenswrapper[4796]: W0930 16:28:06.241155 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d32d14f_94c5_4714_823a_8875a3d840db.slice/crio-ea01ac037e422085448760eb75cbd15fc7cef23d4cfd7f42aad0d5987d497a67 WatchSource:0}: Error finding container ea01ac037e422085448760eb75cbd15fc7cef23d4cfd7f42aad0d5987d497a67: Status 404 returned error can't find the container with id ea01ac037e422085448760eb75cbd15fc7cef23d4cfd7f42aad0d5987d497a67 Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.415430 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.417040 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.421014 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n2dwc" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.421066 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.421351 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.421630 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.433555 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.444214 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.463230 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.478658 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487618 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487679 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-logs\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487706 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487721 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487755 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487782 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487818 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.487887 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmgwz\" (UniqueName: \"kubernetes.io/projected/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-kube-api-access-zmgwz\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: E0930 16:28:06.503393 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b39aff1c-79f0-4747-a111-7c5ac3a7dc3f" containerName="mariadb-account-create" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.503435 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b39aff1c-79f0-4747-a111-7c5ac3a7dc3f" containerName="mariadb-account-create" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.503786 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b39aff1c-79f0-4747-a111-7c5ac3a7dc3f" containerName="mariadb-account-create" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.505413 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.516320 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.524272 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.557608 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589399 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmgwz\" (UniqueName: \"kubernetes.io/projected/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-kube-api-access-zmgwz\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589591 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589655 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-logs\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589678 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589704 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589755 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.589822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.591169 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.593915 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-logs\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.595610 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.612018 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.635738 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.650806 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmgwz\" (UniqueName: \"kubernetes.io/projected/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-kube-api-access-zmgwz\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.651289 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.653656 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.666081 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.710830 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8bv6\" (UniqueName: \"kubernetes.io/projected/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f-kube-api-access-l8bv6\") pod \"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f\" (UID: \"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f\") " Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711172 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711201 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711239 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711257 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns2hx\" (UniqueName: \"kubernetes.io/projected/84d98cdd-c809-4021-9f27-d7c0b922a086-kube-api-access-ns2hx\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711293 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-logs\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711320 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711374 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.711402 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.722338 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f-kube-api-access-l8bv6" (OuterVolumeSpecName: "kube-api-access-l8bv6") pod "b39aff1c-79f0-4747-a111-7c5ac3a7dc3f" (UID: "b39aff1c-79f0-4747-a111-7c5ac3a7dc3f"). InnerVolumeSpecName "kube-api-access-l8bv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.730396 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gclst"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.762477 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.812911 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.812950 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813002 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813023 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns2hx\" (UniqueName: \"kubernetes.io/projected/84d98cdd-c809-4021-9f27-d7c0b922a086-kube-api-access-ns2hx\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813060 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-logs\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813086 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813141 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813163 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813207 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8bv6\" (UniqueName: \"kubernetes.io/projected/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f-kube-api-access-l8bv6\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.813461 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.815127 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.815347 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-logs\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.821137 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.823560 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.831091 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.832795 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xp6sc"] Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.833611 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.835600 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns2hx\" (UniqueName: \"kubernetes.io/projected/84d98cdd-c809-4021-9f27-d7c0b922a086-kube-api-access-ns2hx\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:06 crc kubenswrapper[4796]: I0930 16:28:06.852745 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.028771 4796 generic.go:334] "Generic (PLEG): container finished" podID="9b954a1b-4ce8-46ac-bb4e-cac178f07f25" containerID="b748613aaab1be7a7772bce73f5a60b95c6fba290c9edf64eb949149de839be2" exitCode=0 Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.029076 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" event={"ID":"9b954a1b-4ce8-46ac-bb4e-cac178f07f25","Type":"ContainerDied","Data":"b748613aaab1be7a7772bce73f5a60b95c6fba290c9edf64eb949149de839be2"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.029104 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" event={"ID":"9b954a1b-4ce8-46ac-bb4e-cac178f07f25","Type":"ContainerStarted","Data":"a0c7e36772890a83f4699a02b4db631c8eb014fb72b9a3040791c21f3c6b5cfe"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.042797 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-96a6-account-create-xfrx4" event={"ID":"b39aff1c-79f0-4747-a111-7c5ac3a7dc3f","Type":"ContainerDied","Data":"3d9e46ba669c4982462bdfaad6c5106e1af34bdba88725fb98d1a61010991b4b"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.042849 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d9e46ba669c4982462bdfaad6c5106e1af34bdba88725fb98d1a61010991b4b" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.042948 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96a6-account-create-xfrx4" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.049955 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" event={"ID":"165b3d32-fe2a-47a1-8adf-a70222767c2c","Type":"ContainerStarted","Data":"2fcc2d78fd5c6f22da2432c32e526dc8880727ce4263b08e5f8922c02219942f"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.050023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" event={"ID":"165b3d32-fe2a-47a1-8adf-a70222767c2c","Type":"ContainerStarted","Data":"9ca407c3327e1a438a26239778e0320a754578105f14a5fd0a3326cb00b6b9eb"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.052871 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xp6sc" event={"ID":"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164","Type":"ContainerStarted","Data":"026e84746a302fc1e19fc2d155ff886ffbeda25f2d51b69f717770cae79814de"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.057735 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerStarted","Data":"ef03c4f1d42e4855283f85bedd39ab9e6da013e36e4270640fff9794d051990a"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.079592 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerName="dnsmasq-dns" containerID="cri-o://e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc" gracePeriod=10 Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.080686 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bmxm" event={"ID":"3d32d14f-94c5-4714-823a-8875a3d840db","Type":"ContainerStarted","Data":"adf4294f88db5a64ddda486bd1e7f186a709344f4fe70a74a1862529ec398b55"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.080726 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bmxm" event={"ID":"3d32d14f-94c5-4714-823a-8875a3d840db","Type":"ContainerStarted","Data":"ea01ac037e422085448760eb75cbd15fc7cef23d4cfd7f42aad0d5987d497a67"} Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.135433 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7bmxm" podStartSLOduration=2.135411289 podStartE2EDuration="2.135411289s" podCreationTimestamp="2025-09-30 16:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:07.107725697 +0000 UTC m=+979.121004224" watchObservedRunningTime="2025-09-30 16:28:07.135411289 +0000 UTC m=+979.148689826" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.154934 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.418736 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.607103 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.678527 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.732096 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-sb\") pod \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.732494 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-config\") pod \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.732839 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-swift-storage-0\") pod \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.732889 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-svc\") pod \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.732930 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptlps\" (UniqueName: \"kubernetes.io/projected/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-kube-api-access-ptlps\") pod \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.732971 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-nb\") pod \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\" (UID: \"9b954a1b-4ce8-46ac-bb4e-cac178f07f25\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.738312 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-kube-api-access-ptlps" (OuterVolumeSpecName: "kube-api-access-ptlps") pod "9b954a1b-4ce8-46ac-bb4e-cac178f07f25" (UID: "9b954a1b-4ce8-46ac-bb4e-cac178f07f25"). InnerVolumeSpecName "kube-api-access-ptlps". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.759146 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b954a1b-4ce8-46ac-bb4e-cac178f07f25" (UID: "9b954a1b-4ce8-46ac-bb4e-cac178f07f25"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.775513 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b954a1b-4ce8-46ac-bb4e-cac178f07f25" (UID: "9b954a1b-4ce8-46ac-bb4e-cac178f07f25"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.779869 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b954a1b-4ce8-46ac-bb4e-cac178f07f25" (UID: "9b954a1b-4ce8-46ac-bb4e-cac178f07f25"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.786970 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9b954a1b-4ce8-46ac-bb4e-cac178f07f25" (UID: "9b954a1b-4ce8-46ac-bb4e-cac178f07f25"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.794877 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-config" (OuterVolumeSpecName: "config") pod "9b954a1b-4ce8-46ac-bb4e-cac178f07f25" (UID: "9b954a1b-4ce8-46ac-bb4e-cac178f07f25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.834909 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-svc\") pod \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.834991 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr6t6\" (UniqueName: \"kubernetes.io/projected/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-kube-api-access-rr6t6\") pod \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835158 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-swift-storage-0\") pod \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835302 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-nb\") pod \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835333 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-config\") pod \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835355 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-sb\") pod \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\" (UID: \"cd0fbb84-5d18-47bf-a0e9-0193eb580b01\") " Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835849 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835873 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835886 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835897 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835909 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptlps\" (UniqueName: \"kubernetes.io/projected/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-kube-api-access-ptlps\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.835923 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b954a1b-4ce8-46ac-bb4e-cac178f07f25-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.854609 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-kube-api-access-rr6t6" (OuterVolumeSpecName: "kube-api-access-rr6t6") pod "cd0fbb84-5d18-47bf-a0e9-0193eb580b01" (UID: "cd0fbb84-5d18-47bf-a0e9-0193eb580b01"). InnerVolumeSpecName "kube-api-access-rr6t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.884996 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd0fbb84-5d18-47bf-a0e9-0193eb580b01" (UID: "cd0fbb84-5d18-47bf-a0e9-0193eb580b01"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.913242 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-config" (OuterVolumeSpecName: "config") pod "cd0fbb84-5d18-47bf-a0e9-0193eb580b01" (UID: "cd0fbb84-5d18-47bf-a0e9-0193eb580b01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.914069 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd0fbb84-5d18-47bf-a0e9-0193eb580b01" (UID: "cd0fbb84-5d18-47bf-a0e9-0193eb580b01"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.919661 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd0fbb84-5d18-47bf-a0e9-0193eb580b01" (UID: "cd0fbb84-5d18-47bf-a0e9-0193eb580b01"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.930349 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cd0fbb84-5d18-47bf-a0e9-0193eb580b01" (UID: "cd0fbb84-5d18-47bf-a0e9-0193eb580b01"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.939053 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.939110 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr6t6\" (UniqueName: \"kubernetes.io/projected/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-kube-api-access-rr6t6\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.939121 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.939134 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.939145 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.939172 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd0fbb84-5d18-47bf-a0e9-0193eb580b01-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:07 crc kubenswrapper[4796]: I0930 16:28:07.960409 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:07 crc kubenswrapper[4796]: W0930 16:28:07.972721 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d98cdd_c809_4021_9f27_d7c0b922a086.slice/crio-43dc7637821917e87af78aec908404daf8e14a630ae97c8b5773d754c1259f2f WatchSource:0}: Error finding container 43dc7637821917e87af78aec908404daf8e14a630ae97c8b5773d754c1259f2f: Status 404 returned error can't find the container with id 43dc7637821917e87af78aec908404daf8e14a630ae97c8b5773d754c1259f2f Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.090944 4796 generic.go:334] "Generic (PLEG): container finished" podID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerID="e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc" exitCode=0 Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.091044 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" event={"ID":"cd0fbb84-5d18-47bf-a0e9-0193eb580b01","Type":"ContainerDied","Data":"e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc"} Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.091079 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" event={"ID":"cd0fbb84-5d18-47bf-a0e9-0193eb580b01","Type":"ContainerDied","Data":"5fadeae05b52c6e1ac95224c99281f2bd0488351ec67d2761e033a1e5209afee"} Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.091102 4796 scope.go:117] "RemoveContainer" containerID="e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.091267 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-l6lb8" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.099614 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed","Type":"ContainerStarted","Data":"6eaa1492622f02daf18f2d873d37157befad4c4ffb34c276c2eb6d7732225a22"} Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.108272 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.108653 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t" event={"ID":"9b954a1b-4ce8-46ac-bb4e-cac178f07f25","Type":"ContainerDied","Data":"a0c7e36772890a83f4699a02b4db631c8eb014fb72b9a3040791c21f3c6b5cfe"} Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.118377 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d98cdd-c809-4021-9f27-d7c0b922a086","Type":"ContainerStarted","Data":"43dc7637821917e87af78aec908404daf8e14a630ae97c8b5773d754c1259f2f"} Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.119286 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.144690 4796 generic.go:334] "Generic (PLEG): container finished" podID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerID="2fcc2d78fd5c6f22da2432c32e526dc8880727ce4263b08e5f8922c02219942f" exitCode=0 Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.145770 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" event={"ID":"165b3d32-fe2a-47a1-8adf-a70222767c2c","Type":"ContainerDied","Data":"2fcc2d78fd5c6f22da2432c32e526dc8880727ce4263b08e5f8922c02219942f"} Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.163090 4796 scope.go:117] "RemoveContainer" containerID="4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.198691 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.244637 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.339853 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-l6lb8"] Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.344828 4796 scope.go:117] "RemoveContainer" containerID="e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc" Sep 30 16:28:08 crc kubenswrapper[4796]: E0930 16:28:08.346088 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc\": container with ID starting with e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc not found: ID does not exist" containerID="e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.346131 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc"} err="failed to get container status \"e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc\": rpc error: code = NotFound desc = could not find container \"e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc\": container with ID starting with e7182546b8b477bfcf12b42d94b91e596957ffaffa3b34ecd9113ecbefa8c0dc not found: ID does not exist" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.346160 4796 scope.go:117] "RemoveContainer" containerID="4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.349728 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-l6lb8"] Sep 30 16:28:08 crc kubenswrapper[4796]: E0930 16:28:08.350590 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023\": container with ID starting with 4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023 not found: ID does not exist" containerID="4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.350639 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023"} err="failed to get container status \"4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023\": rpc error: code = NotFound desc = could not find container \"4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023\": container with ID starting with 4d278058f1c9a463c09f01a67a23f1019f9fb062fd2f5167f9d6cf772f865023 not found: ID does not exist" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.350668 4796 scope.go:117] "RemoveContainer" containerID="b748613aaab1be7a7772bce73f5a60b95c6fba290c9edf64eb949149de839be2" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.392048 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t"] Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.400650 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tzz7t"] Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.750570 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b954a1b-4ce8-46ac-bb4e-cac178f07f25" path="/var/lib/kubelet/pods/9b954a1b-4ce8-46ac-bb4e-cac178f07f25/volumes" Sep 30 16:28:08 crc kubenswrapper[4796]: I0930 16:28:08.756917 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" path="/var/lib/kubelet/pods/cd0fbb84-5d18-47bf-a0e9-0193eb580b01/volumes" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.155086 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-rtdjm"] Sep 30 16:28:09 crc kubenswrapper[4796]: E0930 16:28:09.155436 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerName="init" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.155448 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerName="init" Sep 30 16:28:09 crc kubenswrapper[4796]: E0930 16:28:09.155460 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerName="dnsmasq-dns" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.155465 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerName="dnsmasq-dns" Sep 30 16:28:09 crc kubenswrapper[4796]: E0930 16:28:09.155481 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b954a1b-4ce8-46ac-bb4e-cac178f07f25" containerName="init" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.155487 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b954a1b-4ce8-46ac-bb4e-cac178f07f25" containerName="init" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.155640 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b954a1b-4ce8-46ac-bb4e-cac178f07f25" containerName="init" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.155653 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd0fbb84-5d18-47bf-a0e9-0193eb580b01" containerName="dnsmasq-dns" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.156188 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.160798 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.169301 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.175373 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kb6jm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.179996 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rtdjm"] Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.198151 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed","Type":"ContainerStarted","Data":"88df117fe6ac4d6085441765caf9f8193cb398e71d36e0611ef2bbe11b8e97f1"} Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.240800 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d98cdd-c809-4021-9f27-d7c0b922a086","Type":"ContainerStarted","Data":"ebd9423ab65c8cc3d1988ddda33ab1c33b2c07f7e6c6911a45f579bc3f58086f"} Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.244841 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" event={"ID":"165b3d32-fe2a-47a1-8adf-a70222767c2c","Type":"ContainerStarted","Data":"67cb8da6d28812b6c5bf57b52974123deefb14ca65ad9d3ad34e5618827f7ca8"} Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.245337 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.281341 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x8nk\" (UniqueName: \"kubernetes.io/projected/4d5f8250-66ae-4368-b6d6-7d0bea34d616-kube-api-access-6x8nk\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.281570 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-db-sync-config-data\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.281667 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d5f8250-66ae-4368-b6d6-7d0bea34d616-etc-machine-id\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.281773 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-config-data\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.281904 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-scripts\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.282033 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-combined-ca-bundle\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.383278 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-combined-ca-bundle\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.383667 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x8nk\" (UniqueName: \"kubernetes.io/projected/4d5f8250-66ae-4368-b6d6-7d0bea34d616-kube-api-access-6x8nk\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.383698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-db-sync-config-data\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.383720 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d5f8250-66ae-4368-b6d6-7d0bea34d616-etc-machine-id\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.383764 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-config-data\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.383844 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-scripts\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.384437 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d5f8250-66ae-4368-b6d6-7d0bea34d616-etc-machine-id\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.388369 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-combined-ca-bundle\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.389964 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-db-sync-config-data\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.395747 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-config-data\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.398707 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-scripts\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.402536 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x8nk\" (UniqueName: \"kubernetes.io/projected/4d5f8250-66ae-4368-b6d6-7d0bea34d616-kube-api-access-6x8nk\") pod \"cinder-db-sync-rtdjm\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.478633 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.977660 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" podStartSLOduration=4.977642425 podStartE2EDuration="4.977642425s" podCreationTimestamp="2025-09-30 16:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:09.266319344 +0000 UTC m=+981.279597881" watchObservedRunningTime="2025-09-30 16:28:09.977642425 +0000 UTC m=+981.990920952" Sep 30 16:28:09 crc kubenswrapper[4796]: I0930 16:28:09.984071 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rtdjm"] Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.258893 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d98cdd-c809-4021-9f27-d7c0b922a086","Type":"ContainerStarted","Data":"2b874eff0b1a0a969b2861a7657b8e7c59147c8c6639e44b2447ec49690a5661"} Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.258996 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-log" containerID="cri-o://ebd9423ab65c8cc3d1988ddda33ab1c33b2c07f7e6c6911a45f579bc3f58086f" gracePeriod=30 Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.259036 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-httpd" containerID="cri-o://2b874eff0b1a0a969b2861a7657b8e7c59147c8c6639e44b2447ec49690a5661" gracePeriod=30 Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.261445 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed","Type":"ContainerStarted","Data":"1bb0fe5e46f252288eeecb33dfca7113d2b0cb4e6ffdf9568d482d4018279d84"} Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.261581 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-log" containerID="cri-o://88df117fe6ac4d6085441765caf9f8193cb398e71d36e0611ef2bbe11b8e97f1" gracePeriod=30 Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.261619 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-httpd" containerID="cri-o://1bb0fe5e46f252288eeecb33dfca7113d2b0cb4e6ffdf9568d482d4018279d84" gracePeriod=30 Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.281612 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.281593951 podStartE2EDuration="5.281593951s" podCreationTimestamp="2025-09-30 16:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:10.280140368 +0000 UTC m=+982.293418895" watchObservedRunningTime="2025-09-30 16:28:10.281593951 +0000 UTC m=+982.294872478" Sep 30 16:28:10 crc kubenswrapper[4796]: I0930 16:28:10.311483 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.311468236 podStartE2EDuration="5.311468236s" podCreationTimestamp="2025-09-30 16:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:10.303550324 +0000 UTC m=+982.316828841" watchObservedRunningTime="2025-09-30 16:28:10.311468236 +0000 UTC m=+982.324746763" Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.273624 4796 generic.go:334] "Generic (PLEG): container finished" podID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerID="2b874eff0b1a0a969b2861a7657b8e7c59147c8c6639e44b2447ec49690a5661" exitCode=0 Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.273662 4796 generic.go:334] "Generic (PLEG): container finished" podID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerID="ebd9423ab65c8cc3d1988ddda33ab1c33b2c07f7e6c6911a45f579bc3f58086f" exitCode=143 Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.278128 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d98cdd-c809-4021-9f27-d7c0b922a086","Type":"ContainerDied","Data":"2b874eff0b1a0a969b2861a7657b8e7c59147c8c6639e44b2447ec49690a5661"} Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.278173 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d98cdd-c809-4021-9f27-d7c0b922a086","Type":"ContainerDied","Data":"ebd9423ab65c8cc3d1988ddda33ab1c33b2c07f7e6c6911a45f579bc3f58086f"} Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.280151 4796 generic.go:334] "Generic (PLEG): container finished" podID="3d32d14f-94c5-4714-823a-8875a3d840db" containerID="adf4294f88db5a64ddda486bd1e7f186a709344f4fe70a74a1862529ec398b55" exitCode=0 Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.280222 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bmxm" event={"ID":"3d32d14f-94c5-4714-823a-8875a3d840db","Type":"ContainerDied","Data":"adf4294f88db5a64ddda486bd1e7f186a709344f4fe70a74a1862529ec398b55"} Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.284337 4796 generic.go:334] "Generic (PLEG): container finished" podID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerID="1bb0fe5e46f252288eeecb33dfca7113d2b0cb4e6ffdf9568d482d4018279d84" exitCode=0 Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.284364 4796 generic.go:334] "Generic (PLEG): container finished" podID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerID="88df117fe6ac4d6085441765caf9f8193cb398e71d36e0611ef2bbe11b8e97f1" exitCode=143 Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.284390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed","Type":"ContainerDied","Data":"1bb0fe5e46f252288eeecb33dfca7113d2b0cb4e6ffdf9568d482d4018279d84"} Sep 30 16:28:11 crc kubenswrapper[4796]: I0930 16:28:11.284417 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed","Type":"ContainerDied","Data":"88df117fe6ac4d6085441765caf9f8193cb398e71d36e0611ef2bbe11b8e97f1"} Sep 30 16:28:12 crc kubenswrapper[4796]: W0930 16:28:12.278412 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d5f8250_66ae_4368_b6d6_7d0bea34d616.slice/crio-0beeda6ca274d5d47936c641aff5d5dcc7b4d506bdfbebe7dff929bedd744512 WatchSource:0}: Error finding container 0beeda6ca274d5d47936c641aff5d5dcc7b4d506bdfbebe7dff929bedd744512: Status 404 returned error can't find the container with id 0beeda6ca274d5d47936c641aff5d5dcc7b4d506bdfbebe7dff929bedd744512 Sep 30 16:28:12 crc kubenswrapper[4796]: I0930 16:28:12.302228 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtdjm" event={"ID":"4d5f8250-66ae-4368-b6d6-7d0bea34d616","Type":"ContainerStarted","Data":"0beeda6ca274d5d47936c641aff5d5dcc7b4d506bdfbebe7dff929bedd744512"} Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.778602 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b5da-account-create-l97vx"] Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.780339 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.783570 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.789137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b5da-account-create-l97vx"] Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.872637 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gqzp\" (UniqueName: \"kubernetes.io/projected/18a102b3-d988-4fed-837c-9217be9c9c2b-kube-api-access-4gqzp\") pod \"barbican-b5da-account-create-l97vx\" (UID: \"18a102b3-d988-4fed-837c-9217be9c9c2b\") " pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.921456 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c1c4-account-create-52985"] Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.922546 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.924380 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.927760 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c1c4-account-create-52985"] Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.974129 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gqzp\" (UniqueName: \"kubernetes.io/projected/18a102b3-d988-4fed-837c-9217be9c9c2b-kube-api-access-4gqzp\") pod \"barbican-b5da-account-create-l97vx\" (UID: \"18a102b3-d988-4fed-837c-9217be9c9c2b\") " pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.974190 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn4sx\" (UniqueName: \"kubernetes.io/projected/604035fc-9731-47c4-b962-329eb8703d89-kube-api-access-rn4sx\") pod \"neutron-c1c4-account-create-52985\" (UID: \"604035fc-9731-47c4-b962-329eb8703d89\") " pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:13 crc kubenswrapper[4796]: I0930 16:28:13.999003 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gqzp\" (UniqueName: \"kubernetes.io/projected/18a102b3-d988-4fed-837c-9217be9c9c2b-kube-api-access-4gqzp\") pod \"barbican-b5da-account-create-l97vx\" (UID: \"18a102b3-d988-4fed-837c-9217be9c9c2b\") " pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.080151 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn4sx\" (UniqueName: \"kubernetes.io/projected/604035fc-9731-47c4-b962-329eb8703d89-kube-api-access-rn4sx\") pod \"neutron-c1c4-account-create-52985\" (UID: \"604035fc-9731-47c4-b962-329eb8703d89\") " pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.096208 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn4sx\" (UniqueName: \"kubernetes.io/projected/604035fc-9731-47c4-b962-329eb8703d89-kube-api-access-rn4sx\") pod \"neutron-c1c4-account-create-52985\" (UID: \"604035fc-9731-47c4-b962-329eb8703d89\") " pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.104699 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.201356 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.239034 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.284531 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-credential-keys\") pod \"3d32d14f-94c5-4714-823a-8875a3d840db\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.284817 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-config-data\") pod \"3d32d14f-94c5-4714-823a-8875a3d840db\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.284892 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nqrr\" (UniqueName: \"kubernetes.io/projected/3d32d14f-94c5-4714-823a-8875a3d840db-kube-api-access-5nqrr\") pod \"3d32d14f-94c5-4714-823a-8875a3d840db\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.284925 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-combined-ca-bundle\") pod \"3d32d14f-94c5-4714-823a-8875a3d840db\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.284961 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-scripts\") pod \"3d32d14f-94c5-4714-823a-8875a3d840db\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.285025 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-fernet-keys\") pod \"3d32d14f-94c5-4714-823a-8875a3d840db\" (UID: \"3d32d14f-94c5-4714-823a-8875a3d840db\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.290844 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3d32d14f-94c5-4714-823a-8875a3d840db" (UID: "3d32d14f-94c5-4714-823a-8875a3d840db"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.291207 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3d32d14f-94c5-4714-823a-8875a3d840db" (UID: "3d32d14f-94c5-4714-823a-8875a3d840db"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.300415 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-scripts" (OuterVolumeSpecName: "scripts") pod "3d32d14f-94c5-4714-823a-8875a3d840db" (UID: "3d32d14f-94c5-4714-823a-8875a3d840db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.302346 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d32d14f-94c5-4714-823a-8875a3d840db-kube-api-access-5nqrr" (OuterVolumeSpecName: "kube-api-access-5nqrr") pod "3d32d14f-94c5-4714-823a-8875a3d840db" (UID: "3d32d14f-94c5-4714-823a-8875a3d840db"). InnerVolumeSpecName "kube-api-access-5nqrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.333408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bmxm" event={"ID":"3d32d14f-94c5-4714-823a-8875a3d840db","Type":"ContainerDied","Data":"ea01ac037e422085448760eb75cbd15fc7cef23d4cfd7f42aad0d5987d497a67"} Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.333452 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea01ac037e422085448760eb75cbd15fc7cef23d4cfd7f42aad0d5987d497a67" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.333450 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bmxm" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.338589 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d32d14f-94c5-4714-823a-8875a3d840db" (UID: "3d32d14f-94c5-4714-823a-8875a3d840db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.340367 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-config-data" (OuterVolumeSpecName: "config-data") pod "3d32d14f-94c5-4714-823a-8875a3d840db" (UID: "3d32d14f-94c5-4714-823a-8875a3d840db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.386886 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.386921 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.386933 4796 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.386943 4796 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.386955 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d32d14f-94c5-4714-823a-8875a3d840db-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.386966 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nqrr\" (UniqueName: \"kubernetes.io/projected/3d32d14f-94c5-4714-823a-8875a3d840db-kube-api-access-5nqrr\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.464053 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494304 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494435 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-config-data\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494490 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmgwz\" (UniqueName: \"kubernetes.io/projected/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-kube-api-access-zmgwz\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494572 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-public-tls-certs\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494629 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-combined-ca-bundle\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494661 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-scripts\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494731 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-logs\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.494765 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-httpd-run\") pod \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\" (UID: \"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed\") " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.496086 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.497078 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-logs" (OuterVolumeSpecName: "logs") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.503198 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-scripts" (OuterVolumeSpecName: "scripts") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.509396 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.518745 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-kube-api-access-zmgwz" (OuterVolumeSpecName: "kube-api-access-zmgwz") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "kube-api-access-zmgwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.540167 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.553464 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-config-data" (OuterVolumeSpecName: "config-data") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.577307 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" (UID: "2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596829 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596880 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596891 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596901 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596909 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596937 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596946 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.596954 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmgwz\" (UniqueName: \"kubernetes.io/projected/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed-kube-api-access-zmgwz\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.614765 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.668116 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b5da-account-create-l97vx"] Sep 30 16:28:14 crc kubenswrapper[4796]: W0930 16:28:14.679537 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18a102b3_d988_4fed_837c_9217be9c9c2b.slice/crio-8c65cb13ab76a1c62650cb92eef5f2f570db31a0c13b6ad3a9c0ce8ade31e783 WatchSource:0}: Error finding container 8c65cb13ab76a1c62650cb92eef5f2f570db31a0c13b6ad3a9c0ce8ade31e783: Status 404 returned error can't find the container with id 8c65cb13ab76a1c62650cb92eef5f2f570db31a0c13b6ad3a9c0ce8ade31e783 Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.699098 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:14 crc kubenswrapper[4796]: I0930 16:28:14.829646 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c1c4-account-create-52985"] Sep 30 16:28:14 crc kubenswrapper[4796]: W0930 16:28:14.838314 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604035fc_9731_47c4_b962_329eb8703d89.slice/crio-a37a6825944c664c1971a795333e8ea849146db79cc0238ae1f2c9b4c13686a9 WatchSource:0}: Error finding container a37a6825944c664c1971a795333e8ea849146db79cc0238ae1f2c9b4c13686a9: Status 404 returned error can't find the container with id a37a6825944c664c1971a795333e8ea849146db79cc0238ae1f2c9b4c13686a9 Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.211586 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311536 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-scripts\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311597 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-internal-tls-certs\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311652 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-combined-ca-bundle\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311688 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-logs\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311733 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311811 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-config-data\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311837 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns2hx\" (UniqueName: \"kubernetes.io/projected/84d98cdd-c809-4021-9f27-d7c0b922a086-kube-api-access-ns2hx\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.311911 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-httpd-run\") pod \"84d98cdd-c809-4021-9f27-d7c0b922a086\" (UID: \"84d98cdd-c809-4021-9f27-d7c0b922a086\") " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.312473 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-logs" (OuterVolumeSpecName: "logs") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.312605 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.338242 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d98cdd-c809-4021-9f27-d7c0b922a086-kube-api-access-ns2hx" (OuterVolumeSpecName: "kube-api-access-ns2hx") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "kube-api-access-ns2hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.338506 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.339168 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-scripts" (OuterVolumeSpecName: "scripts") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.377389 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerStarted","Data":"b4c5f5a1b4f09dbeeb2e92b21a0a522fee57fac5f6eee0d28dfcd180338ec8b2"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.406771 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7bmxm"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.418897 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns2hx\" (UniqueName: \"kubernetes.io/projected/84d98cdd-c809-4021-9f27-d7c0b922a086-kube-api-access-ns2hx\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.418991 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.419006 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.419016 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d98cdd-c809-4021-9f27-d7c0b922a086-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.419039 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.425438 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7bmxm"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.427134 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.436396 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed","Type":"ContainerDied","Data":"6eaa1492622f02daf18f2d873d37157befad4c4ffb34c276c2eb6d7732225a22"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.436459 4796 scope.go:117] "RemoveContainer" containerID="1bb0fe5e46f252288eeecb33dfca7113d2b0cb4e6ffdf9568d482d4018279d84" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.436609 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.441185 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.446664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d98cdd-c809-4021-9f27-d7c0b922a086","Type":"ContainerDied","Data":"43dc7637821917e87af78aec908404daf8e14a630ae97c8b5773d754c1259f2f"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.446768 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.478117 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-config-data" (OuterVolumeSpecName: "config-data") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.478475 4796 generic.go:334] "Generic (PLEG): container finished" podID="18a102b3-d988-4fed-837c-9217be9c9c2b" containerID="65edba31166889b96c3113729099421540f1f89f256dd4f60863d8db753ec64f" exitCode=0 Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.478733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b5da-account-create-l97vx" event={"ID":"18a102b3-d988-4fed-837c-9217be9c9c2b","Type":"ContainerDied","Data":"65edba31166889b96c3113729099421540f1f89f256dd4f60863d8db753ec64f"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.478777 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b5da-account-create-l97vx" event={"ID":"18a102b3-d988-4fed-837c-9217be9c9c2b","Type":"ContainerStarted","Data":"8c65cb13ab76a1c62650cb92eef5f2f570db31a0c13b6ad3a9c0ce8ade31e783"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.503966 4796 scope.go:117] "RemoveContainer" containerID="88df117fe6ac4d6085441765caf9f8193cb398e71d36e0611ef2bbe11b8e97f1" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.505281 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c1c4-account-create-52985" event={"ID":"604035fc-9731-47c4-b962-329eb8703d89","Type":"ContainerStarted","Data":"cdddeec154684c68a56bb45eca5755fac783ede66ea17ac56b59a529131257eb"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.505323 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c1c4-account-create-52985" event={"ID":"604035fc-9731-47c4-b962-329eb8703d89","Type":"ContainerStarted","Data":"a37a6825944c664c1971a795333e8ea849146db79cc0238ae1f2c9b4c13686a9"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.507653 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dnmm6"] Sep 30 16:28:15 crc kubenswrapper[4796]: E0930 16:28:15.511660 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-log" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.511852 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-log" Sep 30 16:28:15 crc kubenswrapper[4796]: E0930 16:28:15.511941 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-httpd" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.512134 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-httpd" Sep 30 16:28:15 crc kubenswrapper[4796]: E0930 16:28:15.512239 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-log" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.512316 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-log" Sep 30 16:28:15 crc kubenswrapper[4796]: E0930 16:28:15.512402 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d32d14f-94c5-4714-823a-8875a3d840db" containerName="keystone-bootstrap" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.512478 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d32d14f-94c5-4714-823a-8875a3d840db" containerName="keystone-bootstrap" Sep 30 16:28:15 crc kubenswrapper[4796]: E0930 16:28:15.519423 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-httpd" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.519661 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-httpd" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.520168 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-httpd" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.520467 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d32d14f-94c5-4714-823a-8875a3d840db" containerName="keystone-bootstrap" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.520559 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" containerName="glance-log" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.520645 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-log" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.520732 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" containerName="glance-httpd" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.534313 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xp6sc" event={"ID":"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164","Type":"ContainerStarted","Data":"957b10dc4d6aa188ce92c1cb6ebdf63d73b274b07cd034d81c222699103b7543"} Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.534398 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.538271 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "84d98cdd-c809-4021-9f27-d7c0b922a086" (UID: "84d98cdd-c809-4021-9f27-d7c0b922a086"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.540840 4796 scope.go:117] "RemoveContainer" containerID="2b874eff0b1a0a969b2861a7657b8e7c59147c8c6639e44b2447ec49690a5661" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.543822 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.543849 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.543858 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.543872 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d98cdd-c809-4021-9f27-d7c0b922a086-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.544499 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dnmm6"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.570441 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxclq" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.570704 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.570837 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.571129 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.578628 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.589362 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.590670 4796 scope.go:117] "RemoveContainer" containerID="ebd9423ab65c8cc3d1988ddda33ab1c33b2c07f7e6c6911a45f579bc3f58086f" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.599947 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.630848 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.636379 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.636529 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.641339 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.643893 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c1c4-account-create-52985" podStartSLOduration=2.643870636 podStartE2EDuration="2.643870636s" podCreationTimestamp="2025-09-30 16:28:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:15.58427193 +0000 UTC m=+987.597550457" watchObservedRunningTime="2025-09-30 16:28:15.643870636 +0000 UTC m=+987.657149153" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.645062 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-combined-ca-bundle\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.645096 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-scripts\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.645124 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrgxd\" (UniqueName: \"kubernetes.io/projected/b07f5969-8f8f-4976-99d8-a484adec2076-kube-api-access-rrgxd\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.645180 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-config-data\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.645202 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-credential-keys\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.645274 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-fernet-keys\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.653452 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xp6sc" podStartSLOduration=3.371987271 podStartE2EDuration="10.653433526s" podCreationTimestamp="2025-09-30 16:28:05 +0000 UTC" firstStartedPulling="2025-09-30 16:28:06.846584706 +0000 UTC m=+978.859863233" lastFinishedPulling="2025-09-30 16:28:14.128030961 +0000 UTC m=+986.141309488" observedRunningTime="2025-09-30 16:28:15.605356378 +0000 UTC m=+987.618634905" watchObservedRunningTime="2025-09-30 16:28:15.653433526 +0000 UTC m=+987.666712063" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746707 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746755 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746775 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746793 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-logs\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746815 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-combined-ca-bundle\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746830 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-scripts\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746856 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrgxd\" (UniqueName: \"kubernetes.io/projected/b07f5969-8f8f-4976-99d8-a484adec2076-kube-api-access-rrgxd\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746900 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-config-data\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746924 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.746950 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.747017 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-credential-keys\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.747071 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.747142 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-fernet-keys\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.747160 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pbff\" (UniqueName: \"kubernetes.io/projected/14ffda24-fedb-45c8-af77-b9776ecc2e3c-kube-api-access-7pbff\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.752392 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-credential-keys\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.753902 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-combined-ca-bundle\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.754479 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-scripts\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.762846 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-fernet-keys\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.763072 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-config-data\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.772932 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrgxd\" (UniqueName: \"kubernetes.io/projected/b07f5969-8f8f-4976-99d8-a484adec2076-kube-api-access-rrgxd\") pod \"keystone-bootstrap-dnmm6\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.803719 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.823133 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.842734 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849263 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849282 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849304 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-logs\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849348 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849364 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849394 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.849443 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pbff\" (UniqueName: \"kubernetes.io/projected/14ffda24-fedb-45c8-af77-b9776ecc2e3c-kube-api-access-7pbff\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.850027 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.851917 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.854108 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-logs\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.854868 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.855486 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.856744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.870169 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.870269 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.875488 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.881759 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.898700 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pbff\" (UniqueName: \"kubernetes.io/projected/14ffda24-fedb-45c8-af77-b9776ecc2e3c-kube-api-access-7pbff\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.907371 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.911772 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.918599 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " pod="openstack/glance-default-external-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953197 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953535 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953608 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953657 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953717 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q8dz\" (UniqueName: \"kubernetes.io/projected/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-kube-api-access-2q8dz\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953789 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.953815 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:15 crc kubenswrapper[4796]: I0930 16:28:15.961830 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.056972 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057038 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057068 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057119 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057152 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057195 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.057229 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q8dz\" (UniqueName: \"kubernetes.io/projected/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-kube-api-access-2q8dz\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.058421 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.059373 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.064033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.065375 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.071659 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.072797 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.074342 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.100082 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.103364 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q8dz\" (UniqueName: \"kubernetes.io/projected/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-kube-api-access-2q8dz\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.112528 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.184120 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sc9nt"] Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.184390 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-sc9nt" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerName="dnsmasq-dns" containerID="cri-o://5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae" gracePeriod=10 Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.207858 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.426794 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dnmm6"] Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.553023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dnmm6" event={"ID":"b07f5969-8f8f-4976-99d8-a484adec2076","Type":"ContainerStarted","Data":"1cc3c51032203096a012b4216634816a1b191ae2e3438751bc4066b293464d4a"} Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.558888 4796 generic.go:334] "Generic (PLEG): container finished" podID="604035fc-9731-47c4-b962-329eb8703d89" containerID="cdddeec154684c68a56bb45eca5755fac783ede66ea17ac56b59a529131257eb" exitCode=0 Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.558955 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c1c4-account-create-52985" event={"ID":"604035fc-9731-47c4-b962-329eb8703d89","Type":"ContainerDied","Data":"cdddeec154684c68a56bb45eca5755fac783ede66ea17ac56b59a529131257eb"} Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.711801 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.745075 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed" path="/var/lib/kubelet/pods/2b8bc0e0-bf6c-4eeb-849a-918bc22e68ed/volumes" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.745757 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d32d14f-94c5-4714-823a-8875a3d840db" path="/var/lib/kubelet/pods/3d32d14f-94c5-4714-823a-8875a3d840db/volumes" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.746812 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d98cdd-c809-4021-9f27-d7c0b922a086" path="/var/lib/kubelet/pods/84d98cdd-c809-4021-9f27-d7c0b922a086/volumes" Sep 30 16:28:16 crc kubenswrapper[4796]: I0930 16:28:16.803586 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:28:17 crc kubenswrapper[4796]: W0930 16:28:17.204848 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14ffda24_fedb_45c8_af77_b9776ecc2e3c.slice/crio-e9abb5ef61384095a83d2e2b096269b1a56decad59f9c2dbfb3026d1373a6cd0 WatchSource:0}: Error finding container e9abb5ef61384095a83d2e2b096269b1a56decad59f9c2dbfb3026d1373a6cd0: Status 404 returned error can't find the container with id e9abb5ef61384095a83d2e2b096269b1a56decad59f9c2dbfb3026d1373a6cd0 Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.424676 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.434395 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.599202 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gqzp\" (UniqueName: \"kubernetes.io/projected/18a102b3-d988-4fed-837c-9217be9c9c2b-kube-api-access-4gqzp\") pod \"18a102b3-d988-4fed-837c-9217be9c9c2b\" (UID: \"18a102b3-d988-4fed-837c-9217be9c9c2b\") " Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.599451 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-sb\") pod \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.599595 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-config\") pod \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.599645 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-nb\") pod \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.599759 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-dns-svc\") pod \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.599809 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw5bp\" (UniqueName: \"kubernetes.io/projected/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-kube-api-access-zw5bp\") pod \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\" (UID: \"f9d104f1-e734-4c39-9e74-f7b39cd9dbde\") " Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.616533 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dnmm6" event={"ID":"b07f5969-8f8f-4976-99d8-a484adec2076","Type":"ContainerStarted","Data":"7961f219c18f25894dfe345b292a8d4746ac437570709ea4cbcc86881889a7e1"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.618743 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a102b3-d988-4fed-837c-9217be9c9c2b-kube-api-access-4gqzp" (OuterVolumeSpecName: "kube-api-access-4gqzp") pod "18a102b3-d988-4fed-837c-9217be9c9c2b" (UID: "18a102b3-d988-4fed-837c-9217be9c9c2b"). InnerVolumeSpecName "kube-api-access-4gqzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.621880 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b5da-account-create-l97vx" event={"ID":"18a102b3-d988-4fed-837c-9217be9c9c2b","Type":"ContainerDied","Data":"8c65cb13ab76a1c62650cb92eef5f2f570db31a0c13b6ad3a9c0ce8ade31e783"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.622005 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c65cb13ab76a1c62650cb92eef5f2f570db31a0c13b6ad3a9c0ce8ade31e783" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.622018 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b5da-account-create-l97vx" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.626330 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-kube-api-access-zw5bp" (OuterVolumeSpecName: "kube-api-access-zw5bp") pod "f9d104f1-e734-4c39-9e74-f7b39cd9dbde" (UID: "f9d104f1-e734-4c39-9e74-f7b39cd9dbde"). InnerVolumeSpecName "kube-api-access-zw5bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.627380 4796 generic.go:334] "Generic (PLEG): container finished" podID="3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" containerID="957b10dc4d6aa188ce92c1cb6ebdf63d73b274b07cd034d81c222699103b7543" exitCode=0 Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.627440 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xp6sc" event={"ID":"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164","Type":"ContainerDied","Data":"957b10dc4d6aa188ce92c1cb6ebdf63d73b274b07cd034d81c222699103b7543"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.635528 4796 generic.go:334] "Generic (PLEG): container finished" podID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerID="5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae" exitCode=0 Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.635654 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sc9nt" event={"ID":"f9d104f1-e734-4c39-9e74-f7b39cd9dbde","Type":"ContainerDied","Data":"5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.635685 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-sc9nt" event={"ID":"f9d104f1-e734-4c39-9e74-f7b39cd9dbde","Type":"ContainerDied","Data":"d0a4357cf73b9a8c89941e2a80cf142b8b696ed3bc981d45e0edd9ea22ac222e"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.635701 4796 scope.go:117] "RemoveContainer" containerID="5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.635821 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-sc9nt" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.642960 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14ffda24-fedb-45c8-af77-b9776ecc2e3c","Type":"ContainerStarted","Data":"e9abb5ef61384095a83d2e2b096269b1a56decad59f9c2dbfb3026d1373a6cd0"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.656612 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dnmm6" podStartSLOduration=2.656591428 podStartE2EDuration="2.656591428s" podCreationTimestamp="2025-09-30 16:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:17.643194496 +0000 UTC m=+989.656473023" watchObservedRunningTime="2025-09-30 16:28:17.656591428 +0000 UTC m=+989.669869955" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.657765 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4","Type":"ContainerStarted","Data":"1f692b1bc147d4def228b6b87a6888e6edc6ba7dd14f912f8445de2a021625dd"} Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.674389 4796 scope.go:117] "RemoveContainer" containerID="1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.682581 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f9d104f1-e734-4c39-9e74-f7b39cd9dbde" (UID: "f9d104f1-e734-4c39-9e74-f7b39cd9dbde"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.690749 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f9d104f1-e734-4c39-9e74-f7b39cd9dbde" (UID: "f9d104f1-e734-4c39-9e74-f7b39cd9dbde"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.698870 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-config" (OuterVolumeSpecName: "config") pod "f9d104f1-e734-4c39-9e74-f7b39cd9dbde" (UID: "f9d104f1-e734-4c39-9e74-f7b39cd9dbde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.708012 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw5bp\" (UniqueName: \"kubernetes.io/projected/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-kube-api-access-zw5bp\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.708044 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gqzp\" (UniqueName: \"kubernetes.io/projected/18a102b3-d988-4fed-837c-9217be9c9c2b-kube-api-access-4gqzp\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.708054 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.708065 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.708075 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.715424 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f9d104f1-e734-4c39-9e74-f7b39cd9dbde" (UID: "f9d104f1-e734-4c39-9e74-f7b39cd9dbde"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.745167 4796 scope.go:117] "RemoveContainer" containerID="5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae" Sep 30 16:28:17 crc kubenswrapper[4796]: E0930 16:28:17.754912 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae\": container with ID starting with 5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae not found: ID does not exist" containerID="5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.754950 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae"} err="failed to get container status \"5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae\": rpc error: code = NotFound desc = could not find container \"5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae\": container with ID starting with 5e4f38fc7cae22bc3c9f4d640bc8a579edaefa565e1a571273e2806c84d05eae not found: ID does not exist" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.754990 4796 scope.go:117] "RemoveContainer" containerID="1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19" Sep 30 16:28:17 crc kubenswrapper[4796]: E0930 16:28:17.755442 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19\": container with ID starting with 1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19 not found: ID does not exist" containerID="1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.755490 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19"} err="failed to get container status \"1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19\": rpc error: code = NotFound desc = could not find container \"1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19\": container with ID starting with 1cdb22924e4f871f7dc2cc26e98485435728717d580c1909d0a3de667a88bc19 not found: ID does not exist" Sep 30 16:28:17 crc kubenswrapper[4796]: I0930 16:28:17.810046 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9d104f1-e734-4c39-9e74-f7b39cd9dbde-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.026239 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sc9nt"] Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.058525 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-sc9nt"] Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.264066 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.419152 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn4sx\" (UniqueName: \"kubernetes.io/projected/604035fc-9731-47c4-b962-329eb8703d89-kube-api-access-rn4sx\") pod \"604035fc-9731-47c4-b962-329eb8703d89\" (UID: \"604035fc-9731-47c4-b962-329eb8703d89\") " Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.423359 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604035fc-9731-47c4-b962-329eb8703d89-kube-api-access-rn4sx" (OuterVolumeSpecName: "kube-api-access-rn4sx") pod "604035fc-9731-47c4-b962-329eb8703d89" (UID: "604035fc-9731-47c4-b962-329eb8703d89"). InnerVolumeSpecName "kube-api-access-rn4sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.521418 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn4sx\" (UniqueName: \"kubernetes.io/projected/604035fc-9731-47c4-b962-329eb8703d89-kube-api-access-rn4sx\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.678193 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4","Type":"ContainerStarted","Data":"7b62ecc64657dc028863ee3ec7626c4bb282ca4474126411be13829493b65274"} Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.683254 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c1c4-account-create-52985" Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.683246 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c1c4-account-create-52985" event={"ID":"604035fc-9731-47c4-b962-329eb8703d89","Type":"ContainerDied","Data":"a37a6825944c664c1971a795333e8ea849146db79cc0238ae1f2c9b4c13686a9"} Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.683391 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a37a6825944c664c1971a795333e8ea849146db79cc0238ae1f2c9b4c13686a9" Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.687203 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerStarted","Data":"d9d19af0e7d2534a88a316ebc51a22c9e4906888ea69917b6a60a8e36c76e2e0"} Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.693499 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14ffda24-fedb-45c8-af77-b9776ecc2e3c","Type":"ContainerStarted","Data":"bc30eee3be60f47c3b1013f77c6ec08b3e60471c3aa80c489a063eb21a6470c2"} Sep 30 16:28:18 crc kubenswrapper[4796]: I0930 16:28:18.744292 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" path="/var/lib/kubelet/pods/f9d104f1-e734-4c39-9e74-f7b39cd9dbde/volumes" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.173634 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-rsljj"] Sep 30 16:28:19 crc kubenswrapper[4796]: E0930 16:28:19.174631 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerName="dnsmasq-dns" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174650 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerName="dnsmasq-dns" Sep 30 16:28:19 crc kubenswrapper[4796]: E0930 16:28:19.174673 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604035fc-9731-47c4-b962-329eb8703d89" containerName="mariadb-account-create" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174679 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="604035fc-9731-47c4-b962-329eb8703d89" containerName="mariadb-account-create" Sep 30 16:28:19 crc kubenswrapper[4796]: E0930 16:28:19.174703 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a102b3-d988-4fed-837c-9217be9c9c2b" containerName="mariadb-account-create" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174712 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a102b3-d988-4fed-837c-9217be9c9c2b" containerName="mariadb-account-create" Sep 30 16:28:19 crc kubenswrapper[4796]: E0930 16:28:19.174724 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerName="init" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174730 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerName="init" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174912 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a102b3-d988-4fed-837c-9217be9c9c2b" containerName="mariadb-account-create" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174937 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="604035fc-9731-47c4-b962-329eb8703d89" containerName="mariadb-account-create" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.174963 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9d104f1-e734-4c39-9e74-f7b39cd9dbde" containerName="dnsmasq-dns" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.175873 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.176080 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.179996 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.183996 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-69zd7" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.184760 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rsljj"] Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.343670 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-config-data\") pod \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.343743 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-logs\") pod \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.343887 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlvgf\" (UniqueName: \"kubernetes.io/projected/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-kube-api-access-mlvgf\") pod \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.343955 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-scripts\") pod \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.344071 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-combined-ca-bundle\") pod \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\" (UID: \"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164\") " Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.344366 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-db-sync-config-data\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.344510 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffxsl\" (UniqueName: \"kubernetes.io/projected/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-kube-api-access-ffxsl\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.344536 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-combined-ca-bundle\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.345297 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-logs" (OuterVolumeSpecName: "logs") pod "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" (UID: "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.350258 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-scripts" (OuterVolumeSpecName: "scripts") pod "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" (UID: "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.364712 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-kube-api-access-mlvgf" (OuterVolumeSpecName: "kube-api-access-mlvgf") pod "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" (UID: "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164"). InnerVolumeSpecName "kube-api-access-mlvgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.373280 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-config-data" (OuterVolumeSpecName: "config-data") pod "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" (UID: "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.434837 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" (UID: "3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.445964 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffxsl\" (UniqueName: \"kubernetes.io/projected/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-kube-api-access-ffxsl\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446034 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-combined-ca-bundle\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446102 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-db-sync-config-data\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446197 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlvgf\" (UniqueName: \"kubernetes.io/projected/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-kube-api-access-mlvgf\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446210 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446220 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446229 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.446238 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.452295 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-db-sync-config-data\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.454519 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-combined-ca-bundle\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.467088 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffxsl\" (UniqueName: \"kubernetes.io/projected/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-kube-api-access-ffxsl\") pod \"barbican-db-sync-rsljj\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.518149 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.704517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4","Type":"ContainerStarted","Data":"697115e46b410831fc44124a45e13f95e4fc2393d6d3f2e6e096e82c6c89c7d1"} Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.707503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xp6sc" event={"ID":"3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164","Type":"ContainerDied","Data":"026e84746a302fc1e19fc2d155ff886ffbeda25f2d51b69f717770cae79814de"} Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.707547 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="026e84746a302fc1e19fc2d155ff886ffbeda25f2d51b69f717770cae79814de" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.707523 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xp6sc" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.710845 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14ffda24-fedb-45c8-af77-b9776ecc2e3c","Type":"ContainerStarted","Data":"f3eb08f83f733b1498e53fc6064a5759d513c70b1eea1ce30ac6c495877461a7"} Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.723274 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.723255721 podStartE2EDuration="4.723255721s" podCreationTimestamp="2025-09-30 16:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:19.720081398 +0000 UTC m=+991.733359925" watchObservedRunningTime="2025-09-30 16:28:19.723255721 +0000 UTC m=+991.736534248" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.753322 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.7532907909999995 podStartE2EDuration="4.753290791s" podCreationTimestamp="2025-09-30 16:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:19.750115688 +0000 UTC m=+991.763394215" watchObservedRunningTime="2025-09-30 16:28:19.753290791 +0000 UTC m=+991.766569318" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.834729 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5785c489dd-mzgjb"] Sep 30 16:28:19 crc kubenswrapper[4796]: E0930 16:28:19.835133 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" containerName="placement-db-sync" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.835149 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" containerName="placement-db-sync" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.835346 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" containerName="placement-db-sync" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.836218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.845429 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.845837 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.846238 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.846382 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.846588 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bv65w" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.862571 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5785c489dd-mzgjb"] Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.954955 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-scripts\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.955370 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-public-tls-certs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.955390 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdkjv\" (UniqueName: \"kubernetes.io/projected/2a8a5505-564d-4173-8f22-869b0bec6e2c-kube-api-access-gdkjv\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.955418 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-combined-ca-bundle\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.955434 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-config-data\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.955496 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-internal-tls-certs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.955520 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a8a5505-564d-4173-8f22-869b0bec6e2c-logs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:19 crc kubenswrapper[4796]: I0930 16:28:19.990380 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rsljj"] Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.056792 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-scripts\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.056871 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-public-tls-certs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.056890 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdkjv\" (UniqueName: \"kubernetes.io/projected/2a8a5505-564d-4173-8f22-869b0bec6e2c-kube-api-access-gdkjv\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.056918 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-combined-ca-bundle\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.056934 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-config-data\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.057000 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-internal-tls-certs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.057017 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a8a5505-564d-4173-8f22-869b0bec6e2c-logs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.061257 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a8a5505-564d-4173-8f22-869b0bec6e2c-logs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.063702 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-public-tls-certs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.064761 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-combined-ca-bundle\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.064796 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-config-data\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.065591 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-scripts\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.067774 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8a5505-564d-4173-8f22-869b0bec6e2c-internal-tls-certs\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.076531 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdkjv\" (UniqueName: \"kubernetes.io/projected/2a8a5505-564d-4173-8f22-869b0bec6e2c-kube-api-access-gdkjv\") pod \"placement-5785c489dd-mzgjb\" (UID: \"2a8a5505-564d-4173-8f22-869b0bec6e2c\") " pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.161885 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.729398 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rsljj" event={"ID":"15717b4f-8c0e-4105-a2da-1fb7c502a4cc","Type":"ContainerStarted","Data":"b28e7bf1e359faeb21d85df739b356114c0d26ccefd4ca6dc804fdfd85d84e1d"} Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.732264 4796 generic.go:334] "Generic (PLEG): container finished" podID="b07f5969-8f8f-4976-99d8-a484adec2076" containerID="7961f219c18f25894dfe345b292a8d4746ac437570709ea4cbcc86881889a7e1" exitCode=0 Sep 30 16:28:20 crc kubenswrapper[4796]: I0930 16:28:20.744490 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dnmm6" event={"ID":"b07f5969-8f8f-4976-99d8-a484adec2076","Type":"ContainerDied","Data":"7961f219c18f25894dfe345b292a8d4746ac437570709ea4cbcc86881889a7e1"} Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.210888 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-dxs8h"] Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.212494 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.214657 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.214768 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4pzz9" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.215642 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.230716 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dxs8h"] Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.335757 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-combined-ca-bundle\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.335795 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-config\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.335814 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwf7f\" (UniqueName: \"kubernetes.io/projected/27218284-a0ab-4297-8d33-d9abc4e1549f-kube-api-access-rwf7f\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.437254 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-combined-ca-bundle\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.437294 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-config\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.437314 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwf7f\" (UniqueName: \"kubernetes.io/projected/27218284-a0ab-4297-8d33-d9abc4e1549f-kube-api-access-rwf7f\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.443652 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-config\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.473939 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwf7f\" (UniqueName: \"kubernetes.io/projected/27218284-a0ab-4297-8d33-d9abc4e1549f-kube-api-access-rwf7f\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.474416 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-combined-ca-bundle\") pod \"neutron-db-sync-dxs8h\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:24 crc kubenswrapper[4796]: I0930 16:28:24.541354 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:25 crc kubenswrapper[4796]: I0930 16:28:25.963282 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 16:28:25 crc kubenswrapper[4796]: I0930 16:28:25.963643 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.001498 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.003173 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.009747 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.065179 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-config-data\") pod \"b07f5969-8f8f-4976-99d8-a484adec2076\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.065352 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-fernet-keys\") pod \"b07f5969-8f8f-4976-99d8-a484adec2076\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.065439 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-combined-ca-bundle\") pod \"b07f5969-8f8f-4976-99d8-a484adec2076\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.065487 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-scripts\") pod \"b07f5969-8f8f-4976-99d8-a484adec2076\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.065524 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrgxd\" (UniqueName: \"kubernetes.io/projected/b07f5969-8f8f-4976-99d8-a484adec2076-kube-api-access-rrgxd\") pod \"b07f5969-8f8f-4976-99d8-a484adec2076\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.065655 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-credential-keys\") pod \"b07f5969-8f8f-4976-99d8-a484adec2076\" (UID: \"b07f5969-8f8f-4976-99d8-a484adec2076\") " Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.110014 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b07f5969-8f8f-4976-99d8-a484adec2076-kube-api-access-rrgxd" (OuterVolumeSpecName: "kube-api-access-rrgxd") pod "b07f5969-8f8f-4976-99d8-a484adec2076" (UID: "b07f5969-8f8f-4976-99d8-a484adec2076"). InnerVolumeSpecName "kube-api-access-rrgxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.115565 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b07f5969-8f8f-4976-99d8-a484adec2076" (UID: "b07f5969-8f8f-4976-99d8-a484adec2076"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.116095 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-scripts" (OuterVolumeSpecName: "scripts") pod "b07f5969-8f8f-4976-99d8-a484adec2076" (UID: "b07f5969-8f8f-4976-99d8-a484adec2076"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.117344 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-config-data" (OuterVolumeSpecName: "config-data") pod "b07f5969-8f8f-4976-99d8-a484adec2076" (UID: "b07f5969-8f8f-4976-99d8-a484adec2076"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.131138 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b07f5969-8f8f-4976-99d8-a484adec2076" (UID: "b07f5969-8f8f-4976-99d8-a484adec2076"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.139417 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b07f5969-8f8f-4976-99d8-a484adec2076" (UID: "b07f5969-8f8f-4976-99d8-a484adec2076"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.168176 4796 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.168205 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.168221 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.168230 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrgxd\" (UniqueName: \"kubernetes.io/projected/b07f5969-8f8f-4976-99d8-a484adec2076-kube-api-access-rrgxd\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.168240 4796 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.168248 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07f5969-8f8f-4976-99d8-a484adec2076-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.209863 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.217472 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.256867 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.274615 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801213 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dnmm6" event={"ID":"b07f5969-8f8f-4976-99d8-a484adec2076","Type":"ContainerDied","Data":"1cc3c51032203096a012b4216634816a1b191ae2e3438751bc4066b293464d4a"} Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801560 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cc3c51032203096a012b4216634816a1b191ae2e3438751bc4066b293464d4a" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801586 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801602 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801615 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801278 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dnmm6" Sep 30 16:28:26 crc kubenswrapper[4796]: I0930 16:28:26.801626 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.139290 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-75d78779f8-w7h2x"] Sep 30 16:28:27 crc kubenswrapper[4796]: E0930 16:28:27.139778 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b07f5969-8f8f-4976-99d8-a484adec2076" containerName="keystone-bootstrap" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.139797 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b07f5969-8f8f-4976-99d8-a484adec2076" containerName="keystone-bootstrap" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.140072 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b07f5969-8f8f-4976-99d8-a484adec2076" containerName="keystone-bootstrap" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.142482 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.148203 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxclq" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.148367 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.148504 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.148787 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.148865 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.151796 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75d78779f8-w7h2x"] Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.187322 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191515 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-internal-tls-certs\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191661 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-combined-ca-bundle\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191720 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-public-tls-certs\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-fernet-keys\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191792 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-credential-keys\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191915 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-scripts\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191951 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw2pr\" (UniqueName: \"kubernetes.io/projected/624ee0f3-fc1d-4e41-aaae-74767b105143-kube-api-access-kw2pr\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.191972 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-config-data\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293516 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-internal-tls-certs\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293556 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-combined-ca-bundle\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-public-tls-certs\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293620 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-fernet-keys\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293664 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-credential-keys\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-scripts\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293728 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw2pr\" (UniqueName: \"kubernetes.io/projected/624ee0f3-fc1d-4e41-aaae-74767b105143-kube-api-access-kw2pr\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.293748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-config-data\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.298351 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-scripts\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.298809 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-fernet-keys\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.298835 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-credential-keys\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.299685 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-config-data\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.302453 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-combined-ca-bundle\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.306290 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-internal-tls-certs\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.313894 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw2pr\" (UniqueName: \"kubernetes.io/projected/624ee0f3-fc1d-4e41-aaae-74767b105143-kube-api-access-kw2pr\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.316714 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/624ee0f3-fc1d-4e41-aaae-74767b105143-public-tls-certs\") pod \"keystone-75d78779f8-w7h2x\" (UID: \"624ee0f3-fc1d-4e41-aaae-74767b105143\") " pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:27 crc kubenswrapper[4796]: I0930 16:28:27.494302 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:28 crc kubenswrapper[4796]: I0930 16:28:28.892303 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:28 crc kubenswrapper[4796]: I0930 16:28:28.892415 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:28:28 crc kubenswrapper[4796]: I0930 16:28:28.900287 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 16:28:28 crc kubenswrapper[4796]: I0930 16:28:28.947421 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 16:28:28 crc kubenswrapper[4796]: I0930 16:28:28.947526 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:28:29 crc kubenswrapper[4796]: I0930 16:28:29.083153 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 16:28:35 crc kubenswrapper[4796]: I0930 16:28:35.302638 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5785c489dd-mzgjb"] Sep 30 16:28:36 crc kubenswrapper[4796]: E0930 16:28:36.027383 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Sep 30 16:28:36 crc kubenswrapper[4796]: E0930 16:28:36.027742 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6x8nk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-rtdjm_openstack(4d5f8250-66ae-4368-b6d6-7d0bea34d616): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 16:28:36 crc kubenswrapper[4796]: E0930 16:28:36.029122 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-rtdjm" podUID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.392701 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dxs8h"] Sep 30 16:28:36 crc kubenswrapper[4796]: W0930 16:28:36.476688 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod624ee0f3_fc1d_4e41_aaae_74767b105143.slice/crio-fd807c08f4b439338669392326aa9343d1e7234805eb78ad6d490608823f0393 WatchSource:0}: Error finding container fd807c08f4b439338669392326aa9343d1e7234805eb78ad6d490608823f0393: Status 404 returned error can't find the container with id fd807c08f4b439338669392326aa9343d1e7234805eb78ad6d490608823f0393 Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.477204 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75d78779f8-w7h2x"] Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.943815 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75d78779f8-w7h2x" event={"ID":"624ee0f3-fc1d-4e41-aaae-74767b105143","Type":"ContainerStarted","Data":"50738c99ba742d784f315a26755f5e0080a1923c4f1dbc8b4cecf7a482255c7c"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.944089 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.944101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75d78779f8-w7h2x" event={"ID":"624ee0f3-fc1d-4e41-aaae-74767b105143","Type":"ContainerStarted","Data":"fd807c08f4b439338669392326aa9343d1e7234805eb78ad6d490608823f0393"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.945535 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dxs8h" event={"ID":"27218284-a0ab-4297-8d33-d9abc4e1549f","Type":"ContainerStarted","Data":"7f110dd7e4ecdf1f0e79d45a53b07abdcb75ca66393e280322686d144aefbc8f"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.945558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dxs8h" event={"ID":"27218284-a0ab-4297-8d33-d9abc4e1549f","Type":"ContainerStarted","Data":"f6a8ace49e49604e1ab389857c236286e48fd65a18420272e117a58a6a52dd62"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.954307 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5785c489dd-mzgjb" event={"ID":"2a8a5505-564d-4173-8f22-869b0bec6e2c","Type":"ContainerStarted","Data":"24f1dab21fe59847dff0bb170184380755449addcf7c20b76c0878fcbb464b0a"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.954350 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5785c489dd-mzgjb" event={"ID":"2a8a5505-564d-4173-8f22-869b0bec6e2c","Type":"ContainerStarted","Data":"e2dc74463a71f2bf399f673e86d6765857ff955b221ec6fbeca212a8dee4058c"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.954361 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5785c489dd-mzgjb" event={"ID":"2a8a5505-564d-4173-8f22-869b0bec6e2c","Type":"ContainerStarted","Data":"dd8631068402ac2d7b7bc52496337ddb9b5609108102088869cc157d8f33d60b"} Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.955076 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.955099 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.964298 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-75d78779f8-w7h2x" podStartSLOduration=9.964280454 podStartE2EDuration="9.964280454s" podCreationTimestamp="2025-09-30 16:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:36.962685468 +0000 UTC m=+1008.975964025" watchObservedRunningTime="2025-09-30 16:28:36.964280454 +0000 UTC m=+1008.977558981" Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.974735 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerStarted","Data":"d4a8bb6f534526df99d4dfedcb1caa6657d1e29ec914573ec1d984b9a6e027fe"} Sep 30 16:28:36 crc kubenswrapper[4796]: E0930 16:28:36.978173 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-rtdjm" podUID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" Sep 30 16:28:36 crc kubenswrapper[4796]: I0930 16:28:36.986757 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5785c489dd-mzgjb" podStartSLOduration=17.986736982 podStartE2EDuration="17.986736982s" podCreationTimestamp="2025-09-30 16:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:36.982686474 +0000 UTC m=+1008.995965001" watchObservedRunningTime="2025-09-30 16:28:36.986736982 +0000 UTC m=+1009.000015509" Sep 30 16:28:37 crc kubenswrapper[4796]: I0930 16:28:37.011531 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-dxs8h" podStartSLOduration=13.011512298 podStartE2EDuration="13.011512298s" podCreationTimestamp="2025-09-30 16:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:37.010272812 +0000 UTC m=+1009.023551339" watchObservedRunningTime="2025-09-30 16:28:37.011512298 +0000 UTC m=+1009.024790815" Sep 30 16:28:42 crc kubenswrapper[4796]: I0930 16:28:42.034624 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rsljj" event={"ID":"15717b4f-8c0e-4105-a2da-1fb7c502a4cc","Type":"ContainerStarted","Data":"6fe006240ef270bbef55215ea4c3cec1191ff6021a600a638b1de4a76384da33"} Sep 30 16:28:42 crc kubenswrapper[4796]: I0930 16:28:42.065618 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-rsljj" podStartSLOduration=2.079253162 podStartE2EDuration="23.065590892s" podCreationTimestamp="2025-09-30 16:28:19 +0000 UTC" firstStartedPulling="2025-09-30 16:28:19.99896862 +0000 UTC m=+992.012247137" lastFinishedPulling="2025-09-30 16:28:40.98530634 +0000 UTC m=+1012.998584867" observedRunningTime="2025-09-30 16:28:42.050453879 +0000 UTC m=+1014.063732446" watchObservedRunningTime="2025-09-30 16:28:42.065590892 +0000 UTC m=+1014.078869459" Sep 30 16:28:45 crc kubenswrapper[4796]: I0930 16:28:45.070480 4796 generic.go:334] "Generic (PLEG): container finished" podID="15717b4f-8c0e-4105-a2da-1fb7c502a4cc" containerID="6fe006240ef270bbef55215ea4c3cec1191ff6021a600a638b1de4a76384da33" exitCode=0 Sep 30 16:28:45 crc kubenswrapper[4796]: I0930 16:28:45.070892 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rsljj" event={"ID":"15717b4f-8c0e-4105-a2da-1fb7c502a4cc","Type":"ContainerDied","Data":"6fe006240ef270bbef55215ea4c3cec1191ff6021a600a638b1de4a76384da33"} Sep 30 16:28:46 crc kubenswrapper[4796]: I0930 16:28:46.884277 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:46 crc kubenswrapper[4796]: I0930 16:28:46.989271 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-db-sync-config-data\") pod \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " Sep 30 16:28:46 crc kubenswrapper[4796]: I0930 16:28:46.989410 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffxsl\" (UniqueName: \"kubernetes.io/projected/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-kube-api-access-ffxsl\") pod \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " Sep 30 16:28:46 crc kubenswrapper[4796]: I0930 16:28:46.989563 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-combined-ca-bundle\") pod \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\" (UID: \"15717b4f-8c0e-4105-a2da-1fb7c502a4cc\") " Sep 30 16:28:46 crc kubenswrapper[4796]: I0930 16:28:46.994615 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "15717b4f-8c0e-4105-a2da-1fb7c502a4cc" (UID: "15717b4f-8c0e-4105-a2da-1fb7c502a4cc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:46 crc kubenswrapper[4796]: I0930 16:28:46.995368 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-kube-api-access-ffxsl" (OuterVolumeSpecName: "kube-api-access-ffxsl") pod "15717b4f-8c0e-4105-a2da-1fb7c502a4cc" (UID: "15717b4f-8c0e-4105-a2da-1fb7c502a4cc"). InnerVolumeSpecName "kube-api-access-ffxsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.016481 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15717b4f-8c0e-4105-a2da-1fb7c502a4cc" (UID: "15717b4f-8c0e-4105-a2da-1fb7c502a4cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.090877 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffxsl\" (UniqueName: \"kubernetes.io/projected/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-kube-api-access-ffxsl\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.091099 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.091160 4796 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15717b4f-8c0e-4105-a2da-1fb7c502a4cc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.093622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rsljj" event={"ID":"15717b4f-8c0e-4105-a2da-1fb7c502a4cc","Type":"ContainerDied","Data":"b28e7bf1e359faeb21d85df739b356114c0d26ccefd4ca6dc804fdfd85d84e1d"} Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.094319 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rsljj" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.093663 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b28e7bf1e359faeb21d85df739b356114c0d26ccefd4ca6dc804fdfd85d84e1d" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.343750 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-8568546bbc-6fvmg"] Sep 30 16:28:47 crc kubenswrapper[4796]: E0930 16:28:47.344463 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15717b4f-8c0e-4105-a2da-1fb7c502a4cc" containerName="barbican-db-sync" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.344540 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="15717b4f-8c0e-4105-a2da-1fb7c502a4cc" containerName="barbican-db-sync" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.344779 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="15717b4f-8c0e-4105-a2da-1fb7c502a4cc" containerName="barbican-db-sync" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.345910 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.350249 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.350608 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-69zd7" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.355384 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.365668 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-8568546bbc-6fvmg"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.390731 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-d5c775bd6-kffv8"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.396442 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d008ca00-46d2-4985-8f9d-3d48774e9a85-logs\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.397464 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xtnj\" (UniqueName: \"kubernetes.io/projected/d008ca00-46d2-4985-8f9d-3d48774e9a85-kube-api-access-8xtnj\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.397621 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-config-data-custom\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.397771 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-combined-ca-bundle\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.398066 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-config-data\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.398809 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.408486 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.414084 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d5c775bd6-kffv8"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.468662 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-ggfgh"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.470418 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.477629 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-ggfgh"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.502764 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xtnj\" (UniqueName: \"kubernetes.io/projected/d008ca00-46d2-4985-8f9d-3d48774e9a85-kube-api-access-8xtnj\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.502816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-config-data-custom\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.502839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-combined-ca-bundle\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.502906 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919b1e9c-ab87-4902-a8e7-cad17ec22883-logs\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.502950 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-config-data\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.502968 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-config-data-custom\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.503018 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-combined-ca-bundle\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.503042 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-config-data\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.503073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hwwb\" (UniqueName: \"kubernetes.io/projected/919b1e9c-ab87-4902-a8e7-cad17ec22883-kube-api-access-2hwwb\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.503095 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d008ca00-46d2-4985-8f9d-3d48774e9a85-logs\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.503498 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d008ca00-46d2-4985-8f9d-3d48774e9a85-logs\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.512354 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-config-data-custom\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.514644 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-combined-ca-bundle\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.520028 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d008ca00-46d2-4985-8f9d-3d48774e9a85-config-data\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.526641 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xtnj\" (UniqueName: \"kubernetes.io/projected/d008ca00-46d2-4985-8f9d-3d48774e9a85-kube-api-access-8xtnj\") pod \"barbican-worker-8568546bbc-6fvmg\" (UID: \"d008ca00-46d2-4985-8f9d-3d48774e9a85\") " pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.526706 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f6db9d768-79ht6"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.528082 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.531657 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.535599 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f6db9d768-79ht6"] Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605050 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-config\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605127 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-config-data\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605151 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605174 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-config-data-custom\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605202 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-combined-ca-bundle\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605247 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hwwb\" (UniqueName: \"kubernetes.io/projected/919b1e9c-ab87-4902-a8e7-cad17ec22883-kube-api-access-2hwwb\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605283 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-combined-ca-bundle\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605304 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm82n\" (UniqueName: \"kubernetes.io/projected/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-kube-api-access-sm82n\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605320 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87760114-a6da-4c5e-b9d9-01dcac41b188-logs\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7n6f\" (UniqueName: \"kubernetes.io/projected/87760114-a6da-4c5e-b9d9-01dcac41b188-kube-api-access-s7n6f\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605364 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605384 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data-custom\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605409 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605440 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605555 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919b1e9c-ab87-4902-a8e7-cad17ec22883-logs\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.605651 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.606048 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919b1e9c-ab87-4902-a8e7-cad17ec22883-logs\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.609580 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-combined-ca-bundle\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.610072 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-config-data-custom\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.610499 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919b1e9c-ab87-4902-a8e7-cad17ec22883-config-data\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.625668 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hwwb\" (UniqueName: \"kubernetes.io/projected/919b1e9c-ab87-4902-a8e7-cad17ec22883-kube-api-access-2hwwb\") pod \"barbican-keystone-listener-d5c775bd6-kffv8\" (UID: \"919b1e9c-ab87-4902-a8e7-cad17ec22883\") " pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.667045 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-8568546bbc-6fvmg" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.707256 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-config\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.708465 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-config\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709187 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709420 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-combined-ca-bundle\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709465 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm82n\" (UniqueName: \"kubernetes.io/projected/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-kube-api-access-sm82n\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709493 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87760114-a6da-4c5e-b9d9-01dcac41b188-logs\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709586 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7n6f\" (UniqueName: \"kubernetes.io/projected/87760114-a6da-4c5e-b9d9-01dcac41b188-kube-api-access-s7n6f\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709670 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data-custom\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709726 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709818 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.709911 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.710710 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87760114-a6da-4c5e-b9d9-01dcac41b188-logs\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.710813 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.711559 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.711757 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.712051 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.714197 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.714544 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-combined-ca-bundle\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.722484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data-custom\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.726282 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7n6f\" (UniqueName: \"kubernetes.io/projected/87760114-a6da-4c5e-b9d9-01dcac41b188-kube-api-access-s7n6f\") pod \"barbican-api-7f6db9d768-79ht6\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.730529 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.733888 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm82n\" (UniqueName: \"kubernetes.io/projected/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-kube-api-access-sm82n\") pod \"dnsmasq-dns-59d5ff467f-ggfgh\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.785438 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:47 crc kubenswrapper[4796]: I0930 16:28:47.867687 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.102943 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerStarted","Data":"a65e8ded6addf1b75382ebd759fcd0b49ea8cf75099408cfeac2811e78b1ab39"} Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.103347 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-central-agent" containerID="cri-o://b4c5f5a1b4f09dbeeb2e92b21a0a522fee57fac5f6eee0d28dfcd180338ec8b2" gracePeriod=30 Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.103621 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.103842 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="proxy-httpd" containerID="cri-o://a65e8ded6addf1b75382ebd759fcd0b49ea8cf75099408cfeac2811e78b1ab39" gracePeriod=30 Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.103885 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="sg-core" containerID="cri-o://d4a8bb6f534526df99d4dfedcb1caa6657d1e29ec914573ec1d984b9a6e027fe" gracePeriod=30 Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.103921 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-notification-agent" containerID="cri-o://d9d19af0e7d2534a88a316ebc51a22c9e4906888ea69917b6a60a8e36c76e2e0" gracePeriod=30 Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.134951 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.68257595 podStartE2EDuration="43.134931573s" podCreationTimestamp="2025-09-30 16:28:05 +0000 UTC" firstStartedPulling="2025-09-30 16:28:06.440100776 +0000 UTC m=+978.453379303" lastFinishedPulling="2025-09-30 16:28:46.892456399 +0000 UTC m=+1018.905734926" observedRunningTime="2025-09-30 16:28:48.128734532 +0000 UTC m=+1020.142013059" watchObservedRunningTime="2025-09-30 16:28:48.134931573 +0000 UTC m=+1020.148210100" Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.151468 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-8568546bbc-6fvmg"] Sep 30 16:28:48 crc kubenswrapper[4796]: W0930 16:28:48.158126 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd008ca00_46d2_4985_8f9d_3d48774e9a85.slice/crio-4c2b1dfe5231fe0043c04a8724f6b298d13ff1d8e307f40c30592731c463a900 WatchSource:0}: Error finding container 4c2b1dfe5231fe0043c04a8724f6b298d13ff1d8e307f40c30592731c463a900: Status 404 returned error can't find the container with id 4c2b1dfe5231fe0043c04a8724f6b298d13ff1d8e307f40c30592731c463a900 Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.316847 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d5c775bd6-kffv8"] Sep 30 16:28:48 crc kubenswrapper[4796]: W0930 16:28:48.333818 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod919b1e9c_ab87_4902_a8e7_cad17ec22883.slice/crio-67258e6e36fb8d12d50e3e107faa495e991f792024ccc15ce1890fe73e6fbdb0 WatchSource:0}: Error finding container 67258e6e36fb8d12d50e3e107faa495e991f792024ccc15ce1890fe73e6fbdb0: Status 404 returned error can't find the container with id 67258e6e36fb8d12d50e3e107faa495e991f792024ccc15ce1890fe73e6fbdb0 Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.337761 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-ggfgh"] Sep 30 16:28:48 crc kubenswrapper[4796]: I0930 16:28:48.523479 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f6db9d768-79ht6"] Sep 30 16:28:48 crc kubenswrapper[4796]: W0930 16:28:48.544258 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87760114_a6da_4c5e_b9d9_01dcac41b188.slice/crio-16b1eae6b0fdeb85a5c9dc9b35f92c7dd0c49c765f20e10cbee776fbed9e0fd6 WatchSource:0}: Error finding container 16b1eae6b0fdeb85a5c9dc9b35f92c7dd0c49c765f20e10cbee776fbed9e0fd6: Status 404 returned error can't find the container with id 16b1eae6b0fdeb85a5c9dc9b35f92c7dd0c49c765f20e10cbee776fbed9e0fd6 Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.114131 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" event={"ID":"919b1e9c-ab87-4902-a8e7-cad17ec22883","Type":"ContainerStarted","Data":"67258e6e36fb8d12d50e3e107faa495e991f792024ccc15ce1890fe73e6fbdb0"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.116387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6db9d768-79ht6" event={"ID":"87760114-a6da-4c5e-b9d9-01dcac41b188","Type":"ContainerStarted","Data":"09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.116424 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6db9d768-79ht6" event={"ID":"87760114-a6da-4c5e-b9d9-01dcac41b188","Type":"ContainerStarted","Data":"8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.116434 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6db9d768-79ht6" event={"ID":"87760114-a6da-4c5e-b9d9-01dcac41b188","Type":"ContainerStarted","Data":"16b1eae6b0fdeb85a5c9dc9b35f92c7dd0c49c765f20e10cbee776fbed9e0fd6"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.117702 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.117736 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.120138 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8568546bbc-6fvmg" event={"ID":"d008ca00-46d2-4985-8f9d-3d48774e9a85","Type":"ContainerStarted","Data":"4c2b1dfe5231fe0043c04a8724f6b298d13ff1d8e307f40c30592731c463a900"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.124368 4796 generic.go:334] "Generic (PLEG): container finished" podID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerID="a65e8ded6addf1b75382ebd759fcd0b49ea8cf75099408cfeac2811e78b1ab39" exitCode=0 Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.124394 4796 generic.go:334] "Generic (PLEG): container finished" podID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerID="d4a8bb6f534526df99d4dfedcb1caa6657d1e29ec914573ec1d984b9a6e027fe" exitCode=2 Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.124403 4796 generic.go:334] "Generic (PLEG): container finished" podID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerID="b4c5f5a1b4f09dbeeb2e92b21a0a522fee57fac5f6eee0d28dfcd180338ec8b2" exitCode=0 Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.124440 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerDied","Data":"a65e8ded6addf1b75382ebd759fcd0b49ea8cf75099408cfeac2811e78b1ab39"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.124462 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerDied","Data":"d4a8bb6f534526df99d4dfedcb1caa6657d1e29ec914573ec1d984b9a6e027fe"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.124472 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerDied","Data":"b4c5f5a1b4f09dbeeb2e92b21a0a522fee57fac5f6eee0d28dfcd180338ec8b2"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.125945 4796 generic.go:334] "Generic (PLEG): container finished" podID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerID="95bb0b1ec52f8af35d6587efccda964b5ef972f9d01f3ee7a2c5ed7247a15a49" exitCode=0 Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.125963 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" event={"ID":"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558","Type":"ContainerDied","Data":"95bb0b1ec52f8af35d6587efccda964b5ef972f9d01f3ee7a2c5ed7247a15a49"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.125993 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" event={"ID":"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558","Type":"ContainerStarted","Data":"32f9444ab28ef5c2a9ab681016260396dbc0306ed2f147fcde22400d2bc5034b"} Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.151664 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f6db9d768-79ht6" podStartSLOduration=2.151642093 podStartE2EDuration="2.151642093s" podCreationTimestamp="2025-09-30 16:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:49.146251675 +0000 UTC m=+1021.159530222" watchObservedRunningTime="2025-09-30 16:28:49.151642093 +0000 UTC m=+1021.164920620" Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.970041 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57cffd5bb6-xzdml"] Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.971659 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.976853 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.977054 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Sep 30 16:28:49 crc kubenswrapper[4796]: I0930 16:28:49.984260 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57cffd5bb6-xzdml"] Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.090656 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-config-data\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.090959 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-internal-tls-certs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.091033 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqv4s\" (UniqueName: \"kubernetes.io/projected/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-kube-api-access-gqv4s\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.091078 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-combined-ca-bundle\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.091109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-public-tls-certs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.091128 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-config-data-custom\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.091232 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-logs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.141179 4796 generic.go:334] "Generic (PLEG): container finished" podID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerID="d9d19af0e7d2534a88a316ebc51a22c9e4906888ea69917b6a60a8e36c76e2e0" exitCode=0 Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.141929 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerDied","Data":"d9d19af0e7d2534a88a316ebc51a22c9e4906888ea69917b6a60a8e36c76e2e0"} Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.193932 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-combined-ca-bundle\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194010 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-public-tls-certs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194039 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-config-data-custom\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-logs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194142 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-config-data\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194167 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-internal-tls-certs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194204 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqv4s\" (UniqueName: \"kubernetes.io/projected/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-kube-api-access-gqv4s\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.194866 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-logs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.211703 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-internal-tls-certs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.211757 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqv4s\" (UniqueName: \"kubernetes.io/projected/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-kube-api-access-gqv4s\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.212714 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-combined-ca-bundle\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.213072 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-config-data-custom\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.215395 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-public-tls-certs\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.220108 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ed35b2-c6fc-4ab6-8054-d0a58f403425-config-data\") pod \"barbican-api-57cffd5bb6-xzdml\" (UID: \"c7ed35b2-c6fc-4ab6-8054-d0a58f403425\") " pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.297584 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398124 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-combined-ca-bundle\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398434 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-config-data\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398483 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-scripts\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398585 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-run-httpd\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398632 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-log-httpd\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398662 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdnjh\" (UniqueName: \"kubernetes.io/projected/8c4f9d73-18f6-4133-87c7-e120155cf81c-kube-api-access-qdnjh\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.398704 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-sg-core-conf-yaml\") pod \"8c4f9d73-18f6-4133-87c7-e120155cf81c\" (UID: \"8c4f9d73-18f6-4133-87c7-e120155cf81c\") " Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.399473 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.400075 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.403121 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-scripts" (OuterVolumeSpecName: "scripts") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.403775 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c4f9d73-18f6-4133-87c7-e120155cf81c-kube-api-access-qdnjh" (OuterVolumeSpecName: "kube-api-access-qdnjh") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "kube-api-access-qdnjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.428785 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.466420 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.485072 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.500837 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.500871 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.500880 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c4f9d73-18f6-4133-87c7-e120155cf81c-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.500890 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdnjh\" (UniqueName: \"kubernetes.io/projected/8c4f9d73-18f6-4133-87c7-e120155cf81c-kube-api-access-qdnjh\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.500899 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.500907 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.513483 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-config-data" (OuterVolumeSpecName: "config-data") pod "8c4f9d73-18f6-4133-87c7-e120155cf81c" (UID: "8c4f9d73-18f6-4133-87c7-e120155cf81c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.607276 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c4f9d73-18f6-4133-87c7-e120155cf81c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:50 crc kubenswrapper[4796]: I0930 16:28:50.893114 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57cffd5bb6-xzdml"] Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.158220 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cffd5bb6-xzdml" event={"ID":"c7ed35b2-c6fc-4ab6-8054-d0a58f403425","Type":"ContainerStarted","Data":"fe7fb3ff646fb699ae7f800a57ab60d0ed3724c18cc8d908e8fb5b344ac9768d"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.158801 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cffd5bb6-xzdml" event={"ID":"c7ed35b2-c6fc-4ab6-8054-d0a58f403425","Type":"ContainerStarted","Data":"10b569c99367bcb972087bf378cee43fd7d16d0f224fb3f9652d115edb4ed801"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.161349 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" event={"ID":"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558","Type":"ContainerStarted","Data":"9cf31073dfbac1047eda027517c439250965b94c52618bf34488bb6d3c7df6d7"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.161517 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.163445 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" event={"ID":"919b1e9c-ab87-4902-a8e7-cad17ec22883","Type":"ContainerStarted","Data":"d24d8e6aaeddadf1920b8b82475205b2654bafd863163827586985030f129170"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.163522 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" event={"ID":"919b1e9c-ab87-4902-a8e7-cad17ec22883","Type":"ContainerStarted","Data":"068a1e09e1fa34662c22c5f70ca2e87393ef11b78fa31a262cba7a6b5cea5397"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.191655 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" podStartSLOduration=4.191633386 podStartE2EDuration="4.191633386s" podCreationTimestamp="2025-09-30 16:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:51.188050601 +0000 UTC m=+1023.201329138" watchObservedRunningTime="2025-09-30 16:28:51.191633386 +0000 UTC m=+1023.204911923" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.223058 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c4f9d73-18f6-4133-87c7-e120155cf81c","Type":"ContainerDied","Data":"ef03c4f1d42e4855283f85bedd39ab9e6da013e36e4270640fff9794d051990a"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.223104 4796 scope.go:117] "RemoveContainer" containerID="a65e8ded6addf1b75382ebd759fcd0b49ea8cf75099408cfeac2811e78b1ab39" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.223232 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.233881 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8568546bbc-6fvmg" event={"ID":"d008ca00-46d2-4985-8f9d-3d48774e9a85","Type":"ContainerStarted","Data":"426c69b5bd99ac60a282ead2ee0e4768882f72b2dba03c3e21b751212deb5a35"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.233933 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8568546bbc-6fvmg" event={"ID":"d008ca00-46d2-4985-8f9d-3d48774e9a85","Type":"ContainerStarted","Data":"01b22b4dbd1e69f8100695d9d9db4d60ce75d11c8c54bcd2559bdd8ec6e50b77"} Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.255725 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-d5c775bd6-kffv8" podStartSLOduration=2.614713111 podStartE2EDuration="4.255709883s" podCreationTimestamp="2025-09-30 16:28:47 +0000 UTC" firstStartedPulling="2025-09-30 16:28:48.336209341 +0000 UTC m=+1020.349487858" lastFinishedPulling="2025-09-30 16:28:49.977206103 +0000 UTC m=+1021.990484630" observedRunningTime="2025-09-30 16:28:51.209287633 +0000 UTC m=+1023.222566160" watchObservedRunningTime="2025-09-30 16:28:51.255709883 +0000 UTC m=+1023.268988410" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.256174 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-8568546bbc-6fvmg" podStartSLOduration=2.486872116 podStartE2EDuration="4.256123845s" podCreationTimestamp="2025-09-30 16:28:47 +0000 UTC" firstStartedPulling="2025-09-30 16:28:48.159795202 +0000 UTC m=+1020.173073729" lastFinishedPulling="2025-09-30 16:28:49.929046931 +0000 UTC m=+1021.942325458" observedRunningTime="2025-09-30 16:28:51.252811558 +0000 UTC m=+1023.266090095" watchObservedRunningTime="2025-09-30 16:28:51.256123845 +0000 UTC m=+1023.269402372" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.283463 4796 scope.go:117] "RemoveContainer" containerID="d4a8bb6f534526df99d4dfedcb1caa6657d1e29ec914573ec1d984b9a6e027fe" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.311571 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.320925 4796 scope.go:117] "RemoveContainer" containerID="d9d19af0e7d2534a88a316ebc51a22c9e4906888ea69917b6a60a8e36c76e2e0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.329773 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.343767 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:51 crc kubenswrapper[4796]: E0930 16:28:51.344229 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-notification-agent" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344250 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-notification-agent" Sep 30 16:28:51 crc kubenswrapper[4796]: E0930 16:28:51.344283 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-central-agent" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344295 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-central-agent" Sep 30 16:28:51 crc kubenswrapper[4796]: E0930 16:28:51.344318 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="sg-core" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344326 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="sg-core" Sep 30 16:28:51 crc kubenswrapper[4796]: E0930 16:28:51.344342 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="proxy-httpd" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344349 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="proxy-httpd" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344578 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="sg-core" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344612 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-notification-agent" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344624 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="ceilometer-central-agent" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.344641 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" containerName="proxy-httpd" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.350173 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.359838 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.363369 4796 scope.go:117] "RemoveContainer" containerID="b4c5f5a1b4f09dbeeb2e92b21a0a522fee57fac5f6eee0d28dfcd180338ec8b2" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.363603 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.363629 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445366 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-scripts\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-run-httpd\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445444 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445581 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-config-data\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlxfl\" (UniqueName: \"kubernetes.io/projected/234bf5be-fcc7-4d70-83d2-4e1fa8850647-kube-api-access-hlxfl\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445638 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.445659 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-log-httpd\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.494215 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.500111 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5785c489dd-mzgjb" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547435 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-config-data\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547491 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlxfl\" (UniqueName: \"kubernetes.io/projected/234bf5be-fcc7-4d70-83d2-4e1fa8850647-kube-api-access-hlxfl\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547557 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547578 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-log-httpd\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547639 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-scripts\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547662 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-run-httpd\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.547708 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.548756 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-run-httpd\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.549031 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-log-httpd\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.554932 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.555523 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-config-data\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.555860 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.556444 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-scripts\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.580009 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlxfl\" (UniqueName: \"kubernetes.io/projected/234bf5be-fcc7-4d70-83d2-4e1fa8850647-kube-api-access-hlxfl\") pod \"ceilometer-0\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " pod="openstack/ceilometer-0" Sep 30 16:28:51 crc kubenswrapper[4796]: I0930 16:28:51.694535 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.244368 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtdjm" event={"ID":"4d5f8250-66ae-4368-b6d6-7d0bea34d616","Type":"ContainerStarted","Data":"7237daed360dc265a07da547ed5916f32dce47c45c71b2d5c28c5b2e59df4ff8"} Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.250633 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cffd5bb6-xzdml" event={"ID":"c7ed35b2-c6fc-4ab6-8054-d0a58f403425","Type":"ContainerStarted","Data":"eaee014e0b498d87c937344a7709826a379831ed4f6dd1c97190d990e48808ad"} Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.250805 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.250852 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.268609 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-rtdjm" podStartSLOduration=4.426088083 podStartE2EDuration="43.26858773s" podCreationTimestamp="2025-09-30 16:28:09 +0000 UTC" firstStartedPulling="2025-09-30 16:28:12.2816023 +0000 UTC m=+984.294880827" lastFinishedPulling="2025-09-30 16:28:51.124101947 +0000 UTC m=+1023.137380474" observedRunningTime="2025-09-30 16:28:52.264196801 +0000 UTC m=+1024.277475368" watchObservedRunningTime="2025-09-30 16:28:52.26858773 +0000 UTC m=+1024.281866277" Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.297737 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.308131 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57cffd5bb6-xzdml" podStartSLOduration=3.308113268 podStartE2EDuration="3.308113268s" podCreationTimestamp="2025-09-30 16:28:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:52.293541191 +0000 UTC m=+1024.306819718" watchObservedRunningTime="2025-09-30 16:28:52.308113268 +0000 UTC m=+1024.321391785" Sep 30 16:28:52 crc kubenswrapper[4796]: I0930 16:28:52.766635 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c4f9d73-18f6-4133-87c7-e120155cf81c" path="/var/lib/kubelet/pods/8c4f9d73-18f6-4133-87c7-e120155cf81c/volumes" Sep 30 16:28:53 crc kubenswrapper[4796]: I0930 16:28:53.286385 4796 generic.go:334] "Generic (PLEG): container finished" podID="27218284-a0ab-4297-8d33-d9abc4e1549f" containerID="7f110dd7e4ecdf1f0e79d45a53b07abdcb75ca66393e280322686d144aefbc8f" exitCode=0 Sep 30 16:28:53 crc kubenswrapper[4796]: I0930 16:28:53.288697 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dxs8h" event={"ID":"27218284-a0ab-4297-8d33-d9abc4e1549f","Type":"ContainerDied","Data":"7f110dd7e4ecdf1f0e79d45a53b07abdcb75ca66393e280322686d144aefbc8f"} Sep 30 16:28:53 crc kubenswrapper[4796]: I0930 16:28:53.294932 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerStarted","Data":"354c2000822d8b9268fbe152519bcb1b903c31c481790a3925f9d774582c9fb6"} Sep 30 16:28:53 crc kubenswrapper[4796]: I0930 16:28:53.295540 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerStarted","Data":"64220fcd0228cfbd480a312d67127efbbf83f5c250aec547a08d3dd9b7d23c73"} Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.305755 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerStarted","Data":"6c2e1793e1ef97f18659b96ceccd79606e8d5f0614fb4cc5f9c78db24ad7e1c2"} Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.659293 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.827750 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-combined-ca-bundle\") pod \"27218284-a0ab-4297-8d33-d9abc4e1549f\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.827942 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwf7f\" (UniqueName: \"kubernetes.io/projected/27218284-a0ab-4297-8d33-d9abc4e1549f-kube-api-access-rwf7f\") pod \"27218284-a0ab-4297-8d33-d9abc4e1549f\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.828051 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-config\") pod \"27218284-a0ab-4297-8d33-d9abc4e1549f\" (UID: \"27218284-a0ab-4297-8d33-d9abc4e1549f\") " Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.835761 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27218284-a0ab-4297-8d33-d9abc4e1549f-kube-api-access-rwf7f" (OuterVolumeSpecName: "kube-api-access-rwf7f") pod "27218284-a0ab-4297-8d33-d9abc4e1549f" (UID: "27218284-a0ab-4297-8d33-d9abc4e1549f"). InnerVolumeSpecName "kube-api-access-rwf7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.863429 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-config" (OuterVolumeSpecName: "config") pod "27218284-a0ab-4297-8d33-d9abc4e1549f" (UID: "27218284-a0ab-4297-8d33-d9abc4e1549f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.876159 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27218284-a0ab-4297-8d33-d9abc4e1549f" (UID: "27218284-a0ab-4297-8d33-d9abc4e1549f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.930704 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.930744 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27218284-a0ab-4297-8d33-d9abc4e1549f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:54 crc kubenswrapper[4796]: I0930 16:28:54.930762 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwf7f\" (UniqueName: \"kubernetes.io/projected/27218284-a0ab-4297-8d33-d9abc4e1549f-kube-api-access-rwf7f\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.318753 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dxs8h" event={"ID":"27218284-a0ab-4297-8d33-d9abc4e1549f","Type":"ContainerDied","Data":"f6a8ace49e49604e1ab389857c236286e48fd65a18420272e117a58a6a52dd62"} Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.318806 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6a8ace49e49604e1ab389857c236286e48fd65a18420272e117a58a6a52dd62" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.318887 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dxs8h" Sep 30 16:28:55 crc kubenswrapper[4796]: E0930 16:28:55.485487 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27218284_a0ab_4297_8d33_d9abc4e1549f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27218284_a0ab_4297_8d33_d9abc4e1549f.slice/crio-f6a8ace49e49604e1ab389857c236286e48fd65a18420272e117a58a6a52dd62\": RecentStats: unable to find data in memory cache]" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.575582 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-ggfgh"] Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.576094 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerName="dnsmasq-dns" containerID="cri-o://9cf31073dfbac1047eda027517c439250965b94c52618bf34488bb6d3c7df6d7" gracePeriod=10 Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.578373 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.617260 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-6n9vm"] Sep 30 16:28:55 crc kubenswrapper[4796]: E0930 16:28:55.617762 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27218284-a0ab-4297-8d33-d9abc4e1549f" containerName="neutron-db-sync" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.617780 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="27218284-a0ab-4297-8d33-d9abc4e1549f" containerName="neutron-db-sync" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.618051 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="27218284-a0ab-4297-8d33-d9abc4e1549f" containerName="neutron-db-sync" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.619204 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.690238 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-6n9vm"] Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.719961 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77489776dd-pcw55"] Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.721754 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.730473 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.732193 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4pzz9" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.732392 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.732429 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.749648 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-config\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.749899 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.750049 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.750146 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.750217 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.750301 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcbd8\" (UniqueName: \"kubernetes.io/projected/83fac2ba-9d88-4249-9a03-9aebbe20ee54-kube-api-access-vcbd8\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.759968 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77489776dd-pcw55"] Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852275 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhx5x\" (UniqueName: \"kubernetes.io/projected/3df0f7be-028e-4610-9787-14df0c3e5ea0-kube-api-access-qhx5x\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852347 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-combined-ca-bundle\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852378 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-ovndb-tls-certs\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852425 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-config\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852467 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-httpd-config\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852652 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852832 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-config\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852899 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.852963 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.853009 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.853053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcbd8\" (UniqueName: \"kubernetes.io/projected/83fac2ba-9d88-4249-9a03-9aebbe20ee54-kube-api-access-vcbd8\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.853361 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-config\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.854929 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.855005 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.855153 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.855338 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.875120 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcbd8\" (UniqueName: \"kubernetes.io/projected/83fac2ba-9d88-4249-9a03-9aebbe20ee54-kube-api-access-vcbd8\") pod \"dnsmasq-dns-75c8ddd69c-6n9vm\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.954534 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhx5x\" (UniqueName: \"kubernetes.io/projected/3df0f7be-028e-4610-9787-14df0c3e5ea0-kube-api-access-qhx5x\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.954620 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-combined-ca-bundle\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.954658 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-ovndb-tls-certs\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.954713 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-httpd-config\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.954802 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-config\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.959272 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.959491 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-config\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.959744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-httpd-config\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.959752 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-ovndb-tls-certs\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.966766 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-combined-ca-bundle\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:55 crc kubenswrapper[4796]: I0930 16:28:55.979703 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhx5x\" (UniqueName: \"kubernetes.io/projected/3df0f7be-028e-4610-9787-14df0c3e5ea0-kube-api-access-qhx5x\") pod \"neutron-77489776dd-pcw55\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.062238 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.340649 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerStarted","Data":"463499f671252f7a3133a11f7065cd715cfcf89d7667038e15e1449813675ae0"} Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.344111 4796 generic.go:334] "Generic (PLEG): container finished" podID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerID="9cf31073dfbac1047eda027517c439250965b94c52618bf34488bb6d3c7df6d7" exitCode=0 Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.344143 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" event={"ID":"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558","Type":"ContainerDied","Data":"9cf31073dfbac1047eda027517c439250965b94c52618bf34488bb6d3c7df6d7"} Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.516759 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-6n9vm"] Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.671567 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.682544 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77489776dd-pcw55"] Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.773970 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-sb\") pod \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.774148 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-swift-storage-0\") pod \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.774224 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-svc\") pod \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.774267 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm82n\" (UniqueName: \"kubernetes.io/projected/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-kube-api-access-sm82n\") pod \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.774341 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-config\") pod \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.774468 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-nb\") pod \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\" (UID: \"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558\") " Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.782225 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-kube-api-access-sm82n" (OuterVolumeSpecName: "kube-api-access-sm82n") pod "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" (UID: "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558"). InnerVolumeSpecName "kube-api-access-sm82n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.834316 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" (UID: "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.856382 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" (UID: "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.859281 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" (UID: "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.869322 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-config" (OuterVolumeSpecName: "config") pod "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" (UID: "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.874370 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" (UID: "d4b9045a-33bd-4ffd-b9ae-e967fa0e7558"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.882315 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.882347 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.882356 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm82n\" (UniqueName: \"kubernetes.io/projected/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-kube-api-access-sm82n\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.882367 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.882376 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:56 crc kubenswrapper[4796]: I0930 16:28:56.882384 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.355765 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" event={"ID":"d4b9045a-33bd-4ffd-b9ae-e967fa0e7558","Type":"ContainerDied","Data":"32f9444ab28ef5c2a9ab681016260396dbc0306ed2f147fcde22400d2bc5034b"} Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.355836 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-ggfgh" Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.356092 4796 scope.go:117] "RemoveContainer" containerID="9cf31073dfbac1047eda027517c439250965b94c52618bf34488bb6d3c7df6d7" Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.358004 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77489776dd-pcw55" event={"ID":"3df0f7be-028e-4610-9787-14df0c3e5ea0","Type":"ContainerStarted","Data":"823ca9104ebb44966c6cfc1f353cb8d1ccff73e7bb8ffc7deef59bb6be84aa46"} Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.358041 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77489776dd-pcw55" event={"ID":"3df0f7be-028e-4610-9787-14df0c3e5ea0","Type":"ContainerStarted","Data":"0f90adcb820563eb04b8b73ad46757ae6f0643a598f15bb8e359da1f67832b51"} Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.358056 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77489776dd-pcw55" event={"ID":"3df0f7be-028e-4610-9787-14df0c3e5ea0","Type":"ContainerStarted","Data":"ede8dd1c81966e279b43e72afb68ff51827104393f7d620c46534e09d4a267e6"} Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.358085 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.361285 4796 generic.go:334] "Generic (PLEG): container finished" podID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerID="8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425" exitCode=0 Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.361333 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" event={"ID":"83fac2ba-9d88-4249-9a03-9aebbe20ee54","Type":"ContainerDied","Data":"8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425"} Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.361363 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" event={"ID":"83fac2ba-9d88-4249-9a03-9aebbe20ee54","Type":"ContainerStarted","Data":"317269e1b42283d4924acc565814559dd869514a8651308dfb635a9ad79a23ce"} Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.390265 4796 scope.go:117] "RemoveContainer" containerID="95bb0b1ec52f8af35d6587efccda964b5ef972f9d01f3ee7a2c5ed7247a15a49" Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.413075 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77489776dd-pcw55" podStartSLOduration=2.413047563 podStartE2EDuration="2.413047563s" podCreationTimestamp="2025-09-30 16:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:57.401246117 +0000 UTC m=+1029.414524644" watchObservedRunningTime="2025-09-30 16:28:57.413047563 +0000 UTC m=+1029.426326090" Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.566357 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-ggfgh"] Sep 30 16:28:57 crc kubenswrapper[4796]: I0930 16:28:57.575345 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-ggfgh"] Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.370627 4796 generic.go:334] "Generic (PLEG): container finished" podID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" containerID="7237daed360dc265a07da547ed5916f32dce47c45c71b2d5c28c5b2e59df4ff8" exitCode=0 Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.370703 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtdjm" event={"ID":"4d5f8250-66ae-4368-b6d6-7d0bea34d616","Type":"ContainerDied","Data":"7237daed360dc265a07da547ed5916f32dce47c45c71b2d5c28c5b2e59df4ff8"} Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.374558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerStarted","Data":"5571610a7f9962b54c27e92b082278d257a59fefe33c06344976f0ceaa8c9c53"} Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.374726 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.376480 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" event={"ID":"83fac2ba-9d88-4249-9a03-9aebbe20ee54","Type":"ContainerStarted","Data":"4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c"} Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.456994 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.977822561 podStartE2EDuration="7.45696482s" podCreationTimestamp="2025-09-30 16:28:51 +0000 UTC" firstStartedPulling="2025-09-30 16:28:52.3016919 +0000 UTC m=+1024.314970427" lastFinishedPulling="2025-09-30 16:28:57.780834159 +0000 UTC m=+1029.794112686" observedRunningTime="2025-09-30 16:28:58.450715887 +0000 UTC m=+1030.463994414" watchObservedRunningTime="2025-09-30 16:28:58.45696482 +0000 UTC m=+1030.470243347" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.475312 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" podStartSLOduration=3.475293707 podStartE2EDuration="3.475293707s" podCreationTimestamp="2025-09-30 16:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:28:58.473668059 +0000 UTC m=+1030.486946586" watchObservedRunningTime="2025-09-30 16:28:58.475293707 +0000 UTC m=+1030.488572234" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.560540 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-685c9578b5-5pzck"] Sep 30 16:28:58 crc kubenswrapper[4796]: E0930 16:28:58.561092 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerName="dnsmasq-dns" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.561107 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerName="dnsmasq-dns" Sep 30 16:28:58 crc kubenswrapper[4796]: E0930 16:28:58.561122 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerName="init" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.561129 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerName="init" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.561317 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" containerName="dnsmasq-dns" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.562212 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.564223 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.575611 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685c9578b5-5pzck"] Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.590095 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.727477 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-httpd-config\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.727536 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-combined-ca-bundle\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.727562 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-ovndb-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.727749 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-internal-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.727897 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-public-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.728034 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-config\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.728169 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8xz4\" (UniqueName: \"kubernetes.io/projected/3d40c196-945d-4f46-90ca-ea6b287a0f7c-kube-api-access-g8xz4\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.750234 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4b9045a-33bd-4ffd-b9ae-e967fa0e7558" path="/var/lib/kubelet/pods/d4b9045a-33bd-4ffd-b9ae-e967fa0e7558/volumes" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.829748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-public-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.829816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-config\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.829878 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8xz4\" (UniqueName: \"kubernetes.io/projected/3d40c196-945d-4f46-90ca-ea6b287a0f7c-kube-api-access-g8xz4\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.829930 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-httpd-config\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.829960 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-combined-ca-bundle\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.829995 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-ovndb-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.830028 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-internal-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.838879 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-ovndb-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.839631 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-combined-ca-bundle\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.841843 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-config\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.848087 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-httpd-config\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.849572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-public-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.856670 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d40c196-945d-4f46-90ca-ea6b287a0f7c-internal-tls-certs\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.862211 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8xz4\" (UniqueName: \"kubernetes.io/projected/3d40c196-945d-4f46-90ca-ea6b287a0f7c-kube-api-access-g8xz4\") pod \"neutron-685c9578b5-5pzck\" (UID: \"3d40c196-945d-4f46-90ca-ea6b287a0f7c\") " pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:58 crc kubenswrapper[4796]: I0930 16:28:58.882885 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.385504 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.458346 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685c9578b5-5pzck"] Sep 30 16:28:59 crc kubenswrapper[4796]: W0930 16:28:59.465194 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d40c196_945d_4f46_90ca_ea6b287a0f7c.slice/crio-9e5039e4725b7028ed412ed84fd2609a29d2c4931a718e6792af769ae1d2b598 WatchSource:0}: Error finding container 9e5039e4725b7028ed412ed84fd2609a29d2c4931a718e6792af769ae1d2b598: Status 404 returned error can't find the container with id 9e5039e4725b7028ed412ed84fd2609a29d2c4931a718e6792af769ae1d2b598 Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.775391 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-75d78779f8-w7h2x" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.778444 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.842547 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.955516 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x8nk\" (UniqueName: \"kubernetes.io/projected/4d5f8250-66ae-4368-b6d6-7d0bea34d616-kube-api-access-6x8nk\") pod \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.955870 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-scripts\") pod \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.955924 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-combined-ca-bundle\") pod \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.955990 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-config-data\") pod \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.956017 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d5f8250-66ae-4368-b6d6-7d0bea34d616-etc-machine-id\") pod \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.956071 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-db-sync-config-data\") pod \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\" (UID: \"4d5f8250-66ae-4368-b6d6-7d0bea34d616\") " Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.957903 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d5f8250-66ae-4368-b6d6-7d0bea34d616-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4d5f8250-66ae-4368-b6d6-7d0bea34d616" (UID: "4d5f8250-66ae-4368-b6d6-7d0bea34d616"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.966531 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d5f8250-66ae-4368-b6d6-7d0bea34d616-kube-api-access-6x8nk" (OuterVolumeSpecName: "kube-api-access-6x8nk") pod "4d5f8250-66ae-4368-b6d6-7d0bea34d616" (UID: "4d5f8250-66ae-4368-b6d6-7d0bea34d616"). InnerVolumeSpecName "kube-api-access-6x8nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.977085 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-scripts" (OuterVolumeSpecName: "scripts") pod "4d5f8250-66ae-4368-b6d6-7d0bea34d616" (UID: "4d5f8250-66ae-4368-b6d6-7d0bea34d616"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:28:59 crc kubenswrapper[4796]: I0930 16:28:59.981119 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4d5f8250-66ae-4368-b6d6-7d0bea34d616" (UID: "4d5f8250-66ae-4368-b6d6-7d0bea34d616"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.004192 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d5f8250-66ae-4368-b6d6-7d0bea34d616" (UID: "4d5f8250-66ae-4368-b6d6-7d0bea34d616"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.058580 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.058809 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.058912 4796 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d5f8250-66ae-4368-b6d6-7d0bea34d616-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.059501 4796 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.059596 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x8nk\" (UniqueName: \"kubernetes.io/projected/4d5f8250-66ae-4368-b6d6-7d0bea34d616-kube-api-access-6x8nk\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.060113 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-config-data" (OuterVolumeSpecName: "config-data") pod "4d5f8250-66ae-4368-b6d6-7d0bea34d616" (UID: "4d5f8250-66ae-4368-b6d6-7d0bea34d616"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.160957 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d5f8250-66ae-4368-b6d6-7d0bea34d616-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.318895 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.404584 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtdjm" event={"ID":"4d5f8250-66ae-4368-b6d6-7d0bea34d616","Type":"ContainerDied","Data":"0beeda6ca274d5d47936c641aff5d5dcc7b4d506bdfbebe7dff929bedd744512"} Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.404636 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0beeda6ca274d5d47936c641aff5d5dcc7b4d506bdfbebe7dff929bedd744512" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.404692 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtdjm" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.426052 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685c9578b5-5pzck" event={"ID":"3d40c196-945d-4f46-90ca-ea6b287a0f7c","Type":"ContainerStarted","Data":"e3e31b30725819123506b3e8cba6a4ae25849a38af03f324be27b8dfe4a0f8ee"} Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.426097 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685c9578b5-5pzck" event={"ID":"3d40c196-945d-4f46-90ca-ea6b287a0f7c","Type":"ContainerStarted","Data":"bf68d148a6616b712118968095012ea642eb42e21583d07d0a7c636211285754"} Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.426117 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.426130 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685c9578b5-5pzck" event={"ID":"3d40c196-945d-4f46-90ca-ea6b287a0f7c","Type":"ContainerStarted","Data":"9e5039e4725b7028ed412ed84fd2609a29d2c4931a718e6792af769ae1d2b598"} Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.465582 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-685c9578b5-5pzck" podStartSLOduration=2.465565221 podStartE2EDuration="2.465565221s" podCreationTimestamp="2025-09-30 16:28:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:00.450072858 +0000 UTC m=+1032.463351385" watchObservedRunningTime="2025-09-30 16:29:00.465565221 +0000 UTC m=+1032.478843748" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.572897 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:00 crc kubenswrapper[4796]: E0930 16:29:00.577433 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" containerName="cinder-db-sync" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.577470 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" containerName="cinder-db-sync" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.577710 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" containerName="cinder-db-sync" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.578916 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.583853 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.584020 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kb6jm" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.584194 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.584289 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.603761 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.675911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c5td\" (UniqueName: \"kubernetes.io/projected/7bd58138-cb97-4641-b73f-aa3abb1aec45-kube-api-access-2c5td\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.675987 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.676143 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.676179 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bd58138-cb97-4641-b73f-aa3abb1aec45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.676199 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-scripts\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.676281 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.704493 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-6n9vm"] Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.729145 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nnxbc"] Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.741189 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.778621 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nnxbc"] Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.870029 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.872744 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.871965 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-config\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873524 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-svc\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873546 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873611 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873719 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c5td\" (UniqueName: \"kubernetes.io/projected/7bd58138-cb97-4641-b73f-aa3abb1aec45-kube-api-access-2c5td\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873741 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873810 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873843 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873920 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czskd\" (UniqueName: \"kubernetes.io/projected/26bce871-dd26-4faf-a8a4-772a16ae7df2-kube-api-access-czskd\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873941 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bd58138-cb97-4641-b73f-aa3abb1aec45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.873968 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-scripts\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.879130 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.881187 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.874965 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bd58138-cb97-4641-b73f-aa3abb1aec45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.875462 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.886791 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-scripts\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.898551 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.900962 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.928822 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c5td\" (UniqueName: \"kubernetes.io/projected/7bd58138-cb97-4641-b73f-aa3abb1aec45-kube-api-access-2c5td\") pod \"cinder-scheduler-0\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:00 crc kubenswrapper[4796]: I0930 16:29:00.950582 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.038489 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.038628 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.038808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czskd\" (UniqueName: \"kubernetes.io/projected/26bce871-dd26-4faf-a8a4-772a16ae7df2-kube-api-access-czskd\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.038915 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-logs\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039082 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjjfv\" (UniqueName: \"kubernetes.io/projected/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-kube-api-access-zjjfv\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039171 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039215 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039353 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039449 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-config\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039612 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-svc\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039677 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039719 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.039876 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-scripts\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.041510 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.042954 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.043550 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-config\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.043654 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-svc\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.044268 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.054848 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czskd\" (UniqueName: \"kubernetes.io/projected/26bce871-dd26-4faf-a8a4-772a16ae7df2-kube-api-access-czskd\") pod \"dnsmasq-dns-5784cf869f-nnxbc\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.067753 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.141880 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-logs\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.141933 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjjfv\" (UniqueName: \"kubernetes.io/projected/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-kube-api-access-zjjfv\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.142358 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-logs\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.143827 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.143966 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.144079 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.144120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-scripts\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.144153 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.144241 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.148613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-scripts\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.148770 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.149218 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.167283 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.175921 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjjfv\" (UniqueName: \"kubernetes.io/projected/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-kube-api-access-zjjfv\") pod \"cinder-api-0\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.220262 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.319711 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.487118 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerName="dnsmasq-dns" containerID="cri-o://4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c" gracePeriod=10 Sep 30 16:29:01 crc kubenswrapper[4796]: I0930 16:29:01.660810 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nnxbc"] Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.067262 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.170721 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.244493 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.395821 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-sb\") pod \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.395968 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcbd8\" (UniqueName: \"kubernetes.io/projected/83fac2ba-9d88-4249-9a03-9aebbe20ee54-kube-api-access-vcbd8\") pod \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.396050 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-swift-storage-0\") pod \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.396082 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-nb\") pod \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.396128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-config\") pod \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.396235 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-svc\") pod \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\" (UID: \"83fac2ba-9d88-4249-9a03-9aebbe20ee54\") " Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.411199 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83fac2ba-9d88-4249-9a03-9aebbe20ee54-kube-api-access-vcbd8" (OuterVolumeSpecName: "kube-api-access-vcbd8") pod "83fac2ba-9d88-4249-9a03-9aebbe20ee54" (UID: "83fac2ba-9d88-4249-9a03-9aebbe20ee54"). InnerVolumeSpecName "kube-api-access-vcbd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.463928 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83fac2ba-9d88-4249-9a03-9aebbe20ee54" (UID: "83fac2ba-9d88-4249-9a03-9aebbe20ee54"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.498126 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcbd8\" (UniqueName: \"kubernetes.io/projected/83fac2ba-9d88-4249-9a03-9aebbe20ee54-kube-api-access-vcbd8\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.498412 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.525522 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bd58138-cb97-4641-b73f-aa3abb1aec45","Type":"ContainerStarted","Data":"01dc280c64a8b4c887f76805ad18b5d9fe5729b1d926af1dde2fcc4039ec4e2e"} Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.527632 4796 generic.go:334] "Generic (PLEG): container finished" podID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerID="981589e343b339b655a98d3e8e67ab8e93106995711ce81b59076ab7126c3454" exitCode=0 Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.528767 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" event={"ID":"26bce871-dd26-4faf-a8a4-772a16ae7df2","Type":"ContainerDied","Data":"981589e343b339b655a98d3e8e67ab8e93106995711ce81b59076ab7126c3454"} Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.528788 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" event={"ID":"26bce871-dd26-4faf-a8a4-772a16ae7df2","Type":"ContainerStarted","Data":"6ec5467dd9031a6cdc415c942a44bda1b566f8f78edcbc67f20e9138d7b2df4d"} Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.533232 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83fac2ba-9d88-4249-9a03-9aebbe20ee54" (UID: "83fac2ba-9d88-4249-9a03-9aebbe20ee54"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.556226 4796 generic.go:334] "Generic (PLEG): container finished" podID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerID="4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c" exitCode=0 Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.556308 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" event={"ID":"83fac2ba-9d88-4249-9a03-9aebbe20ee54","Type":"ContainerDied","Data":"4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c"} Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.556338 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" event={"ID":"83fac2ba-9d88-4249-9a03-9aebbe20ee54","Type":"ContainerDied","Data":"317269e1b42283d4924acc565814559dd869514a8651308dfb635a9ad79a23ce"} Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.556358 4796 scope.go:117] "RemoveContainer" containerID="4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.556481 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-6n9vm" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.577289 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8","Type":"ContainerStarted","Data":"c0311f8e903d7f7727143319f269b655f376c0455d30ff3712ad51bea0e5eea5"} Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.579395 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-config" (OuterVolumeSpecName: "config") pod "83fac2ba-9d88-4249-9a03-9aebbe20ee54" (UID: "83fac2ba-9d88-4249-9a03-9aebbe20ee54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.602421 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.602463 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.700309 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "83fac2ba-9d88-4249-9a03-9aebbe20ee54" (UID: "83fac2ba-9d88-4249-9a03-9aebbe20ee54"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.704721 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.715930 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "83fac2ba-9d88-4249-9a03-9aebbe20ee54" (UID: "83fac2ba-9d88-4249-9a03-9aebbe20ee54"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.813128 4796 scope.go:117] "RemoveContainer" containerID="8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.813244 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83fac2ba-9d88-4249-9a03-9aebbe20ee54-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.851148 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 30 16:29:02 crc kubenswrapper[4796]: E0930 16:29:02.851512 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerName="init" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.851525 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerName="init" Sep 30 16:29:02 crc kubenswrapper[4796]: E0930 16:29:02.851547 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerName="dnsmasq-dns" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.851553 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerName="dnsmasq-dns" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.851752 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" containerName="dnsmasq-dns" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.852412 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.862256 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rzbrz" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.862473 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.862617 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.884445 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.984587 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-6n9vm"] Sep 30 16:29:02 crc kubenswrapper[4796]: I0930 16:29:02.984632 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-6n9vm"] Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.018192 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config-secret\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.018512 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.018544 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjmwg\" (UniqueName: \"kubernetes.io/projected/411c488d-d370-41c8-93e1-13c6721eb41d-kube-api-access-mjmwg\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.018620 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.121187 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.121313 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config-secret\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.121347 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.121364 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjmwg\" (UniqueName: \"kubernetes.io/projected/411c488d-d370-41c8-93e1-13c6721eb41d-kube-api-access-mjmwg\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.123776 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.126371 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config-secret\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.128811 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.143284 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjmwg\" (UniqueName: \"kubernetes.io/projected/411c488d-d370-41c8-93e1-13c6721eb41d-kube-api-access-mjmwg\") pod \"openstackclient\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.238440 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.249378 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.249455 4796 scope.go:117] "RemoveContainer" containerID="4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c" Sep 30 16:29:03 crc kubenswrapper[4796]: E0930 16:29:03.251360 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c\": container with ID starting with 4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c not found: ID does not exist" containerID="4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.251403 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c"} err="failed to get container status \"4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c\": rpc error: code = NotFound desc = could not find container \"4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c\": container with ID starting with 4a8cf03b2fe26b30b42d66839da621b5478a22afd274e091775299a7dfaf789c not found: ID does not exist" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.251429 4796 scope.go:117] "RemoveContainer" containerID="8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425" Sep 30 16:29:03 crc kubenswrapper[4796]: E0930 16:29:03.256192 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425\": container with ID starting with 8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425 not found: ID does not exist" containerID="8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.256234 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425"} err="failed to get container status \"8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425\": rpc error: code = NotFound desc = could not find container \"8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425\": container with ID starting with 8a7f5724c176a462ede945c48948348bed6120d9515bc91f8ec69db8f4609425 not found: ID does not exist" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.357873 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.649504 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" event={"ID":"26bce871-dd26-4faf-a8a4-772a16ae7df2","Type":"ContainerStarted","Data":"9839668dab97df9c5c115feb6c77cbd558717eb50a56cb16e7ad2cf2747940fc"} Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.649837 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.663186 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57cffd5bb6-xzdml" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.681276 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8","Type":"ContainerStarted","Data":"8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06"} Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.701747 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" podStartSLOduration=3.701731174 podStartE2EDuration="3.701731174s" podCreationTimestamp="2025-09-30 16:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:03.679571275 +0000 UTC m=+1035.692849802" watchObservedRunningTime="2025-09-30 16:29:03.701731174 +0000 UTC m=+1035.715009701" Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.759465 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f6db9d768-79ht6"] Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.759858 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f6db9d768-79ht6" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api-log" containerID="cri-o://8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16" gracePeriod=30 Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.760084 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f6db9d768-79ht6" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api" containerID="cri-o://09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd" gracePeriod=30 Sep 30 16:29:03 crc kubenswrapper[4796]: I0930 16:29:03.852491 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.705111 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bd58138-cb97-4641-b73f-aa3abb1aec45","Type":"ContainerStarted","Data":"95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1"} Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.728733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8","Type":"ContainerStarted","Data":"e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6"} Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.728955 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api-log" containerID="cri-o://8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06" gracePeriod=30 Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.729209 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api" containerID="cri-o://e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6" gracePeriod=30 Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.729416 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.749130 4796 generic.go:334] "Generic (PLEG): container finished" podID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerID="8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16" exitCode=143 Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.757682 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83fac2ba-9d88-4249-9a03-9aebbe20ee54" path="/var/lib/kubelet/pods/83fac2ba-9d88-4249-9a03-9aebbe20ee54/volumes" Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.758335 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6db9d768-79ht6" event={"ID":"87760114-a6da-4c5e-b9d9-01dcac41b188","Type":"ContainerDied","Data":"8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16"} Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.767569 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.767544008 podStartE2EDuration="4.767544008s" podCreationTimestamp="2025-09-30 16:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:04.748374257 +0000 UTC m=+1036.761652784" watchObservedRunningTime="2025-09-30 16:29:04.767544008 +0000 UTC m=+1036.780822535" Sep 30 16:29:04 crc kubenswrapper[4796]: I0930 16:29:04.768205 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"411c488d-d370-41c8-93e1-13c6721eb41d","Type":"ContainerStarted","Data":"160ae577cee63cefcd23811997fd7a47d0609784824295722b9a5621c7ada3cb"} Sep 30 16:29:05 crc kubenswrapper[4796]: I0930 16:29:05.788760 4796 generic.go:334] "Generic (PLEG): container finished" podID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerID="8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06" exitCode=143 Sep 30 16:29:05 crc kubenswrapper[4796]: I0930 16:29:05.789117 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8","Type":"ContainerDied","Data":"8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06"} Sep 30 16:29:05 crc kubenswrapper[4796]: I0930 16:29:05.793074 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bd58138-cb97-4641-b73f-aa3abb1aec45","Type":"ContainerStarted","Data":"c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b"} Sep 30 16:29:05 crc kubenswrapper[4796]: I0930 16:29:05.817184 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.597004913 podStartE2EDuration="5.817167728s" podCreationTimestamp="2025-09-30 16:29:00 +0000 UTC" firstStartedPulling="2025-09-30 16:29:02.094606063 +0000 UTC m=+1034.107884590" lastFinishedPulling="2025-09-30 16:29:03.314768878 +0000 UTC m=+1035.328047405" observedRunningTime="2025-09-30 16:29:05.813271214 +0000 UTC m=+1037.826549741" watchObservedRunningTime="2025-09-30 16:29:05.817167728 +0000 UTC m=+1037.830446255" Sep 30 16:29:06 crc kubenswrapper[4796]: I0930 16:29:06.221573 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.751184 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.818491 4796 generic.go:334] "Generic (PLEG): container finished" podID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerID="09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd" exitCode=0 Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.818577 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f6db9d768-79ht6" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.818582 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6db9d768-79ht6" event={"ID":"87760114-a6da-4c5e-b9d9-01dcac41b188","Type":"ContainerDied","Data":"09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd"} Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.818630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6db9d768-79ht6" event={"ID":"87760114-a6da-4c5e-b9d9-01dcac41b188","Type":"ContainerDied","Data":"16b1eae6b0fdeb85a5c9dc9b35f92c7dd0c49c765f20e10cbee776fbed9e0fd6"} Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.818770 4796 scope.go:117] "RemoveContainer" containerID="09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.849331 4796 scope.go:117] "RemoveContainer" containerID="8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.853061 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data-custom\") pod \"87760114-a6da-4c5e-b9d9-01dcac41b188\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.853099 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87760114-a6da-4c5e-b9d9-01dcac41b188-logs\") pod \"87760114-a6da-4c5e-b9d9-01dcac41b188\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.853155 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7n6f\" (UniqueName: \"kubernetes.io/projected/87760114-a6da-4c5e-b9d9-01dcac41b188-kube-api-access-s7n6f\") pod \"87760114-a6da-4c5e-b9d9-01dcac41b188\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.853184 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data\") pod \"87760114-a6da-4c5e-b9d9-01dcac41b188\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.853272 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-combined-ca-bundle\") pod \"87760114-a6da-4c5e-b9d9-01dcac41b188\" (UID: \"87760114-a6da-4c5e-b9d9-01dcac41b188\") " Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.854895 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87760114-a6da-4c5e-b9d9-01dcac41b188-logs" (OuterVolumeSpecName: "logs") pod "87760114-a6da-4c5e-b9d9-01dcac41b188" (UID: "87760114-a6da-4c5e-b9d9-01dcac41b188"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.859596 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "87760114-a6da-4c5e-b9d9-01dcac41b188" (UID: "87760114-a6da-4c5e-b9d9-01dcac41b188"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.861087 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87760114-a6da-4c5e-b9d9-01dcac41b188-kube-api-access-s7n6f" (OuterVolumeSpecName: "kube-api-access-s7n6f") pod "87760114-a6da-4c5e-b9d9-01dcac41b188" (UID: "87760114-a6da-4c5e-b9d9-01dcac41b188"). InnerVolumeSpecName "kube-api-access-s7n6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.881523 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87760114-a6da-4c5e-b9d9-01dcac41b188" (UID: "87760114-a6da-4c5e-b9d9-01dcac41b188"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.886424 4796 scope.go:117] "RemoveContainer" containerID="09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd" Sep 30 16:29:07 crc kubenswrapper[4796]: E0930 16:29:07.886956 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd\": container with ID starting with 09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd not found: ID does not exist" containerID="09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.887014 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd"} err="failed to get container status \"09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd\": rpc error: code = NotFound desc = could not find container \"09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd\": container with ID starting with 09446d75dc106347067bc2a6c71d4194d31f0389bb570da01252aaa9c70b43bd not found: ID does not exist" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.887041 4796 scope.go:117] "RemoveContainer" containerID="8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16" Sep 30 16:29:07 crc kubenswrapper[4796]: E0930 16:29:07.887472 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16\": container with ID starting with 8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16 not found: ID does not exist" containerID="8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.887505 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16"} err="failed to get container status \"8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16\": rpc error: code = NotFound desc = could not find container \"8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16\": container with ID starting with 8e7bac59a883dd6c26dc24cdce178c0f67ed3261adf098d7e5ff50f5ba859a16 not found: ID does not exist" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.921506 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data" (OuterVolumeSpecName: "config-data") pod "87760114-a6da-4c5e-b9d9-01dcac41b188" (UID: "87760114-a6da-4c5e-b9d9-01dcac41b188"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.955907 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.955944 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87760114-a6da-4c5e-b9d9-01dcac41b188-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.955954 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7n6f\" (UniqueName: \"kubernetes.io/projected/87760114-a6da-4c5e-b9d9-01dcac41b188-kube-api-access-s7n6f\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.955965 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:07 crc kubenswrapper[4796]: I0930 16:29:07.955973 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87760114-a6da-4c5e-b9d9-01dcac41b188-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:08 crc kubenswrapper[4796]: I0930 16:29:08.158883 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f6db9d768-79ht6"] Sep 30 16:29:08 crc kubenswrapper[4796]: I0930 16:29:08.166440 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7f6db9d768-79ht6"] Sep 30 16:29:08 crc kubenswrapper[4796]: I0930 16:29:08.769935 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" path="/var/lib/kubelet/pods/87760114-a6da-4c5e-b9d9-01dcac41b188/volumes" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.708529 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.708813 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-central-agent" containerID="cri-o://354c2000822d8b9268fbe152519bcb1b903c31c481790a3925f9d774582c9fb6" gracePeriod=30 Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.709496 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="proxy-httpd" containerID="cri-o://5571610a7f9962b54c27e92b082278d257a59fefe33c06344976f0ceaa8c9c53" gracePeriod=30 Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.709543 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="sg-core" containerID="cri-o://463499f671252f7a3133a11f7065cd715cfcf89d7667038e15e1449813675ae0" gracePeriod=30 Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.709571 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-notification-agent" containerID="cri-o://6c2e1793e1ef97f18659b96ceccd79606e8d5f0614fb4cc5f9c78db24ad7e1c2" gracePeriod=30 Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.722039 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-85f459d7c9-l56gz"] Sep 30 16:29:09 crc kubenswrapper[4796]: E0930 16:29:09.722422 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api-log" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.722438 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api-log" Sep 30 16:29:09 crc kubenswrapper[4796]: E0930 16:29:09.722448 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.722454 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.722649 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api-log" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.722670 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="87760114-a6da-4c5e-b9d9-01dcac41b188" containerName="barbican-api" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.723668 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.725660 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.725924 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.726059 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.729284 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.162:3000/\": EOF" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.734943 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85f459d7c9-l56gz"] Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.786884 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2dj\" (UniqueName: \"kubernetes.io/projected/df49731d-3260-4e0a-8633-e2a7ca68011d-kube-api-access-ld2dj\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787050 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-public-tls-certs\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-combined-ca-bundle\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787092 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df49731d-3260-4e0a-8633-e2a7ca68011d-log-httpd\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787130 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-internal-tls-certs\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787156 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-config-data\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787471 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/df49731d-3260-4e0a-8633-e2a7ca68011d-etc-swift\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.787493 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df49731d-3260-4e0a-8633-e2a7ca68011d-run-httpd\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.847463 4796 generic.go:334] "Generic (PLEG): container finished" podID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerID="463499f671252f7a3133a11f7065cd715cfcf89d7667038e15e1449813675ae0" exitCode=2 Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.847508 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerDied","Data":"463499f671252f7a3133a11f7065cd715cfcf89d7667038e15e1449813675ae0"} Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.888773 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-public-tls-certs\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.888827 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-combined-ca-bundle\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.888853 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df49731d-3260-4e0a-8633-e2a7ca68011d-log-httpd\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.888890 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-internal-tls-certs\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.888947 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-config-data\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.889021 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/df49731d-3260-4e0a-8633-e2a7ca68011d-etc-swift\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.889039 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df49731d-3260-4e0a-8633-e2a7ca68011d-run-httpd\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.889087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2dj\" (UniqueName: \"kubernetes.io/projected/df49731d-3260-4e0a-8633-e2a7ca68011d-kube-api-access-ld2dj\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.889828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df49731d-3260-4e0a-8633-e2a7ca68011d-log-httpd\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.889901 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df49731d-3260-4e0a-8633-e2a7ca68011d-run-httpd\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.895171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-public-tls-certs\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.895918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-combined-ca-bundle\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.896208 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-internal-tls-certs\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.896902 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df49731d-3260-4e0a-8633-e2a7ca68011d-config-data\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.899467 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/df49731d-3260-4e0a-8633-e2a7ca68011d-etc-swift\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:09 crc kubenswrapper[4796]: I0930 16:29:09.905753 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2dj\" (UniqueName: \"kubernetes.io/projected/df49731d-3260-4e0a-8633-e2a7ca68011d-kube-api-access-ld2dj\") pod \"swift-proxy-85f459d7c9-l56gz\" (UID: \"df49731d-3260-4e0a-8633-e2a7ca68011d\") " pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:10 crc kubenswrapper[4796]: I0930 16:29:10.059038 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:10 crc kubenswrapper[4796]: I0930 16:29:10.859152 4796 generic.go:334] "Generic (PLEG): container finished" podID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerID="5571610a7f9962b54c27e92b082278d257a59fefe33c06344976f0ceaa8c9c53" exitCode=0 Sep 30 16:29:10 crc kubenswrapper[4796]: I0930 16:29:10.859479 4796 generic.go:334] "Generic (PLEG): container finished" podID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerID="354c2000822d8b9268fbe152519bcb1b903c31c481790a3925f9d774582c9fb6" exitCode=0 Sep 30 16:29:10 crc kubenswrapper[4796]: I0930 16:29:10.859229 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerDied","Data":"5571610a7f9962b54c27e92b082278d257a59fefe33c06344976f0ceaa8c9c53"} Sep 30 16:29:10 crc kubenswrapper[4796]: I0930 16:29:10.859521 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerDied","Data":"354c2000822d8b9268fbe152519bcb1b903c31c481790a3925f9d774582c9fb6"} Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.069731 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.121314 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gclst"] Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.121908 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerName="dnsmasq-dns" containerID="cri-o://67cb8da6d28812b6c5bf57b52974123deefb14ca65ad9d3ad34e5618827f7ca8" gracePeriod=10 Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.500137 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.537772 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.869161 4796 generic.go:334] "Generic (PLEG): container finished" podID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerID="67cb8da6d28812b6c5bf57b52974123deefb14ca65ad9d3ad34e5618827f7ca8" exitCode=0 Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.869676 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="cinder-scheduler" containerID="cri-o://95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1" gracePeriod=30 Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.869233 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" event={"ID":"165b3d32-fe2a-47a1-8adf-a70222767c2c","Type":"ContainerDied","Data":"67cb8da6d28812b6c5bf57b52974123deefb14ca65ad9d3ad34e5618827f7ca8"} Sep 30 16:29:11 crc kubenswrapper[4796]: I0930 16:29:11.870092 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="probe" containerID="cri-o://c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b" gracePeriod=30 Sep 30 16:29:12 crc kubenswrapper[4796]: I0930 16:29:12.880496 4796 generic.go:334] "Generic (PLEG): container finished" podID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerID="c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b" exitCode=0 Sep 30 16:29:12 crc kubenswrapper[4796]: I0930 16:29:12.880543 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bd58138-cb97-4641-b73f-aa3abb1aec45","Type":"ContainerDied","Data":"c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b"} Sep 30 16:29:13 crc kubenswrapper[4796]: I0930 16:29:13.181129 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 16:29:13 crc kubenswrapper[4796]: I0930 16:29:13.927474 4796 generic.go:334] "Generic (PLEG): container finished" podID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerID="6c2e1793e1ef97f18659b96ceccd79606e8d5f0614fb4cc5f9c78db24ad7e1c2" exitCode=0 Sep 30 16:29:13 crc kubenswrapper[4796]: I0930 16:29:13.927508 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerDied","Data":"6c2e1793e1ef97f18659b96ceccd79606e8d5f0614fb4cc5f9c78db24ad7e1c2"} Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.875964 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888543 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-sg-core-conf-yaml\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888609 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlxfl\" (UniqueName: \"kubernetes.io/projected/234bf5be-fcc7-4d70-83d2-4e1fa8850647-kube-api-access-hlxfl\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888643 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-log-httpd\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888731 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-scripts\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888764 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-combined-ca-bundle\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888822 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-config-data\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.888952 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-run-httpd\") pod \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\" (UID: \"234bf5be-fcc7-4d70-83d2-4e1fa8850647\") " Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.890265 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.891621 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.900507 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-scripts" (OuterVolumeSpecName: "scripts") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.901147 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234bf5be-fcc7-4d70-83d2-4e1fa8850647-kube-api-access-hlxfl" (OuterVolumeSpecName: "kube-api-access-hlxfl") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "kube-api-access-hlxfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.926578 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.938209 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"411c488d-d370-41c8-93e1-13c6721eb41d","Type":"ContainerStarted","Data":"f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86"} Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.943823 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"234bf5be-fcc7-4d70-83d2-4e1fa8850647","Type":"ContainerDied","Data":"64220fcd0228cfbd480a312d67127efbbf83f5c250aec547a08d3dd9b7d23c73"} Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.943876 4796 scope.go:117] "RemoveContainer" containerID="5571610a7f9962b54c27e92b082278d257a59fefe33c06344976f0ceaa8c9c53" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.944069 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.954618 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.958842 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.217268922 podStartE2EDuration="12.958826281s" podCreationTimestamp="2025-09-30 16:29:02 +0000 UTC" firstStartedPulling="2025-09-30 16:29:03.891378284 +0000 UTC m=+1035.904656811" lastFinishedPulling="2025-09-30 16:29:14.632935643 +0000 UTC m=+1046.646214170" observedRunningTime="2025-09-30 16:29:14.953773603 +0000 UTC m=+1046.967052130" watchObservedRunningTime="2025-09-30 16:29:14.958826281 +0000 UTC m=+1046.972104828" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.985375 4796 scope.go:117] "RemoveContainer" containerID="463499f671252f7a3133a11f7065cd715cfcf89d7667038e15e1449813675ae0" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.991746 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.991775 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.991788 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlxfl\" (UniqueName: \"kubernetes.io/projected/234bf5be-fcc7-4d70-83d2-4e1fa8850647-kube-api-access-hlxfl\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.991800 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/234bf5be-fcc7-4d70-83d2-4e1fa8850647-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:14 crc kubenswrapper[4796]: I0930 16:29:14.991811 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.016914 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.019062 4796 scope.go:117] "RemoveContainer" containerID="6c2e1793e1ef97f18659b96ceccd79606e8d5f0614fb4cc5f9c78db24ad7e1c2" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.038745 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-config-data" (OuterVolumeSpecName: "config-data") pod "234bf5be-fcc7-4d70-83d2-4e1fa8850647" (UID: "234bf5be-fcc7-4d70-83d2-4e1fa8850647"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.044152 4796 scope.go:117] "RemoveContainer" containerID="354c2000822d8b9268fbe152519bcb1b903c31c481790a3925f9d774582c9fb6" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.094684 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-config\") pod \"165b3d32-fe2a-47a1-8adf-a70222767c2c\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.094836 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-swift-storage-0\") pod \"165b3d32-fe2a-47a1-8adf-a70222767c2c\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.094887 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-svc\") pod \"165b3d32-fe2a-47a1-8adf-a70222767c2c\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.094921 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-nb\") pod \"165b3d32-fe2a-47a1-8adf-a70222767c2c\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.095034 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-sb\") pod \"165b3d32-fe2a-47a1-8adf-a70222767c2c\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.095087 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxgdq\" (UniqueName: \"kubernetes.io/projected/165b3d32-fe2a-47a1-8adf-a70222767c2c-kube-api-access-nxgdq\") pod \"165b3d32-fe2a-47a1-8adf-a70222767c2c\" (UID: \"165b3d32-fe2a-47a1-8adf-a70222767c2c\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.095579 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.095601 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234bf5be-fcc7-4d70-83d2-4e1fa8850647-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.103201 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/165b3d32-fe2a-47a1-8adf-a70222767c2c-kube-api-access-nxgdq" (OuterVolumeSpecName: "kube-api-access-nxgdq") pod "165b3d32-fe2a-47a1-8adf-a70222767c2c" (UID: "165b3d32-fe2a-47a1-8adf-a70222767c2c"). InnerVolumeSpecName "kube-api-access-nxgdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.150502 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "165b3d32-fe2a-47a1-8adf-a70222767c2c" (UID: "165b3d32-fe2a-47a1-8adf-a70222767c2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.150648 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "165b3d32-fe2a-47a1-8adf-a70222767c2c" (UID: "165b3d32-fe2a-47a1-8adf-a70222767c2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.156121 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85f459d7c9-l56gz"] Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.157718 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "165b3d32-fe2a-47a1-8adf-a70222767c2c" (UID: "165b3d32-fe2a-47a1-8adf-a70222767c2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.160855 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-config" (OuterVolumeSpecName: "config") pod "165b3d32-fe2a-47a1-8adf-a70222767c2c" (UID: "165b3d32-fe2a-47a1-8adf-a70222767c2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: W0930 16:29:15.163046 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf49731d_3260_4e0a_8633_e2a7ca68011d.slice/crio-2aa22deee2000d38aedd5d4ddab509aea175cefd622a3813a7052cefff635644 WatchSource:0}: Error finding container 2aa22deee2000d38aedd5d4ddab509aea175cefd622a3813a7052cefff635644: Status 404 returned error can't find the container with id 2aa22deee2000d38aedd5d4ddab509aea175cefd622a3813a7052cefff635644 Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.175257 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "165b3d32-fe2a-47a1-8adf-a70222767c2c" (UID: "165b3d32-fe2a-47a1-8adf-a70222767c2c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.197668 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.197708 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.197725 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.197739 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.197750 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/165b3d32-fe2a-47a1-8adf-a70222767c2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.197761 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxgdq\" (UniqueName: \"kubernetes.io/projected/165b3d32-fe2a-47a1-8adf-a70222767c2c-kube-api-access-nxgdq\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.293629 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.306533 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.322730 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:15 crc kubenswrapper[4796]: E0930 16:29:15.323312 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="proxy-httpd" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.323392 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="proxy-httpd" Sep 30 16:29:15 crc kubenswrapper[4796]: E0930 16:29:15.323447 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="sg-core" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.323505 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="sg-core" Sep 30 16:29:15 crc kubenswrapper[4796]: E0930 16:29:15.323571 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-central-agent" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.323620 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-central-agent" Sep 30 16:29:15 crc kubenswrapper[4796]: E0930 16:29:15.323676 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-notification-agent" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.323733 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-notification-agent" Sep 30 16:29:15 crc kubenswrapper[4796]: E0930 16:29:15.323789 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerName="dnsmasq-dns" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.323837 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerName="dnsmasq-dns" Sep 30 16:29:15 crc kubenswrapper[4796]: E0930 16:29:15.323902 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerName="init" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.323952 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerName="init" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.324183 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-notification-agent" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.324261 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="sg-core" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.324326 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" containerName="dnsmasq-dns" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.324387 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="ceilometer-central-agent" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.324438 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" containerName="proxy-httpd" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.326031 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.328324 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.328845 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.337831 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.402910 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-run-httpd\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.402972 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xp6r\" (UniqueName: \"kubernetes.io/projected/93081c0f-7233-4958-be78-a3cadc41d385-kube-api-access-4xp6r\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.403004 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-log-httpd\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.403042 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-scripts\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.403098 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.403115 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.403137 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-config-data\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.505769 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-run-httpd\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.505857 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xp6r\" (UniqueName: \"kubernetes.io/projected/93081c0f-7233-4958-be78-a3cadc41d385-kube-api-access-4xp6r\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.505902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-log-httpd\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.506265 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-run-httpd\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.505935 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-scripts\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.507272 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.507310 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.507336 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-config-data\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.508010 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-log-httpd\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.512181 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.512971 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-scripts\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.513077 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-config-data\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.514510 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.523764 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xp6r\" (UniqueName: \"kubernetes.io/projected/93081c0f-7233-4958-be78-a3cadc41d385-kube-api-access-4xp6r\") pod \"ceilometer-0\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.644610 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.815912 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.929679 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data-custom\") pod \"7bd58138-cb97-4641-b73f-aa3abb1aec45\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.930352 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data\") pod \"7bd58138-cb97-4641-b73f-aa3abb1aec45\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.930477 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-scripts\") pod \"7bd58138-cb97-4641-b73f-aa3abb1aec45\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.930507 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bd58138-cb97-4641-b73f-aa3abb1aec45-etc-machine-id\") pod \"7bd58138-cb97-4641-b73f-aa3abb1aec45\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.930530 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-combined-ca-bundle\") pod \"7bd58138-cb97-4641-b73f-aa3abb1aec45\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.930551 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c5td\" (UniqueName: \"kubernetes.io/projected/7bd58138-cb97-4641-b73f-aa3abb1aec45-kube-api-access-2c5td\") pod \"7bd58138-cb97-4641-b73f-aa3abb1aec45\" (UID: \"7bd58138-cb97-4641-b73f-aa3abb1aec45\") " Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.934049 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bd58138-cb97-4641-b73f-aa3abb1aec45-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7bd58138-cb97-4641-b73f-aa3abb1aec45" (UID: "7bd58138-cb97-4641-b73f-aa3abb1aec45"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.939354 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-scripts" (OuterVolumeSpecName: "scripts") pod "7bd58138-cb97-4641-b73f-aa3abb1aec45" (UID: "7bd58138-cb97-4641-b73f-aa3abb1aec45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.945132 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bd58138-cb97-4641-b73f-aa3abb1aec45-kube-api-access-2c5td" (OuterVolumeSpecName: "kube-api-access-2c5td") pod "7bd58138-cb97-4641-b73f-aa3abb1aec45" (UID: "7bd58138-cb97-4641-b73f-aa3abb1aec45"). InnerVolumeSpecName "kube-api-access-2c5td". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.946054 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7bd58138-cb97-4641-b73f-aa3abb1aec45" (UID: "7bd58138-cb97-4641-b73f-aa3abb1aec45"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.974940 4796 generic.go:334] "Generic (PLEG): container finished" podID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerID="95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1" exitCode=0 Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.975071 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bd58138-cb97-4641-b73f-aa3abb1aec45","Type":"ContainerDied","Data":"95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1"} Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.975101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bd58138-cb97-4641-b73f-aa3abb1aec45","Type":"ContainerDied","Data":"01dc280c64a8b4c887f76805ad18b5d9fe5729b1d926af1dde2fcc4039ec4e2e"} Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.975117 4796 scope.go:117] "RemoveContainer" containerID="c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.975408 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.983494 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" event={"ID":"165b3d32-fe2a-47a1-8adf-a70222767c2c","Type":"ContainerDied","Data":"9ca407c3327e1a438a26239778e0320a754578105f14a5fd0a3326cb00b6b9eb"} Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.983907 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gclst" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.992068 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85f459d7c9-l56gz" event={"ID":"df49731d-3260-4e0a-8633-e2a7ca68011d","Type":"ContainerStarted","Data":"6d94aabac1d73f19e99a3f163ff04bc60878aede8316512f231b5b0942a5666b"} Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.992106 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85f459d7c9-l56gz" event={"ID":"df49731d-3260-4e0a-8633-e2a7ca68011d","Type":"ContainerStarted","Data":"1d7d689a40171249169078cc98e85cc995e856abd72c25c44c625f1ef8caf0f5"} Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.992116 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85f459d7c9-l56gz" event={"ID":"df49731d-3260-4e0a-8633-e2a7ca68011d","Type":"ContainerStarted","Data":"2aa22deee2000d38aedd5d4ddab509aea175cefd622a3813a7052cefff635644"} Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.992138 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:15 crc kubenswrapper[4796]: I0930 16:29:15.992151 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.006297 4796 scope.go:117] "RemoveContainer" containerID="95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.016048 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-85f459d7c9-l56gz" podStartSLOduration=7.016026833 podStartE2EDuration="7.016026833s" podCreationTimestamp="2025-09-30 16:29:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:16.010923054 +0000 UTC m=+1048.024201601" watchObservedRunningTime="2025-09-30 16:29:16.016026833 +0000 UTC m=+1048.029305360" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.032148 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.032176 4796 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bd58138-cb97-4641-b73f-aa3abb1aec45-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.032187 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c5td\" (UniqueName: \"kubernetes.io/projected/7bd58138-cb97-4641-b73f-aa3abb1aec45-kube-api-access-2c5td\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.032198 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.052644 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gclst"] Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.063868 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bd58138-cb97-4641-b73f-aa3abb1aec45" (UID: "7bd58138-cb97-4641-b73f-aa3abb1aec45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.068549 4796 scope.go:117] "RemoveContainer" containerID="c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b" Sep 30 16:29:16 crc kubenswrapper[4796]: E0930 16:29:16.069860 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b\": container with ID starting with c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b not found: ID does not exist" containerID="c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.069925 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b"} err="failed to get container status \"c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b\": rpc error: code = NotFound desc = could not find container \"c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b\": container with ID starting with c45f1bb0929756d79d0a7b276453bf1b10a4424a808c0810a80ac5cd16c5d90b not found: ID does not exist" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.069957 4796 scope.go:117] "RemoveContainer" containerID="95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1" Sep 30 16:29:16 crc kubenswrapper[4796]: E0930 16:29:16.075118 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1\": container with ID starting with 95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1 not found: ID does not exist" containerID="95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.075349 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1"} err="failed to get container status \"95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1\": rpc error: code = NotFound desc = could not find container \"95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1\": container with ID starting with 95c250e778313ce7e5b9cf15763ead55152a7c814924150a40da71af32b516f1 not found: ID does not exist" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.075380 4796 scope.go:117] "RemoveContainer" containerID="67cb8da6d28812b6c5bf57b52974123deefb14ca65ad9d3ad34e5618827f7ca8" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.079708 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gclst"] Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.102100 4796 scope.go:117] "RemoveContainer" containerID="2fcc2d78fd5c6f22da2432c32e526dc8880727ce4263b08e5f8922c02219942f" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.133840 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.137273 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data" (OuterVolumeSpecName: "config-data") pod "7bd58138-cb97-4641-b73f-aa3abb1aec45" (UID: "7bd58138-cb97-4641-b73f-aa3abb1aec45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.184301 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.235871 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd58138-cb97-4641-b73f-aa3abb1aec45-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.309860 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.316832 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.333769 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:16 crc kubenswrapper[4796]: E0930 16:29:16.334226 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="cinder-scheduler" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.334244 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="cinder-scheduler" Sep 30 16:29:16 crc kubenswrapper[4796]: E0930 16:29:16.334268 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="probe" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.334274 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="probe" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.334451 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="probe" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.334483 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" containerName="cinder-scheduler" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.335522 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.340020 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.366517 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.439401 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.439446 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxcln\" (UniqueName: \"kubernetes.io/projected/cf5d97fd-5337-41e8-9480-86b0342e9598-kube-api-access-vxcln\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.439470 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-scripts\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.439564 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.439592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-config-data\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.439615 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf5d97fd-5337-41e8-9480-86b0342e9598-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541226 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541300 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-config-data\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541333 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf5d97fd-5337-41e8-9480-86b0342e9598-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541412 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541444 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxcln\" (UniqueName: \"kubernetes.io/projected/cf5d97fd-5337-41e8-9480-86b0342e9598-kube-api-access-vxcln\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541502 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-scripts\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.541443 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf5d97fd-5337-41e8-9480-86b0342e9598-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.548215 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.548496 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-scripts\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.548599 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.552816 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d97fd-5337-41e8-9480-86b0342e9598-config-data\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.562486 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxcln\" (UniqueName: \"kubernetes.io/projected/cf5d97fd-5337-41e8-9480-86b0342e9598-kube-api-access-vxcln\") pod \"cinder-scheduler-0\" (UID: \"cf5d97fd-5337-41e8-9480-86b0342e9598\") " pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.662079 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.752654 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="165b3d32-fe2a-47a1-8adf-a70222767c2c" path="/var/lib/kubelet/pods/165b3d32-fe2a-47a1-8adf-a70222767c2c/volumes" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.753848 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234bf5be-fcc7-4d70-83d2-4e1fa8850647" path="/var/lib/kubelet/pods/234bf5be-fcc7-4d70-83d2-4e1fa8850647/volumes" Sep 30 16:29:16 crc kubenswrapper[4796]: I0930 16:29:16.754795 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bd58138-cb97-4641-b73f-aa3abb1aec45" path="/var/lib/kubelet/pods/7bd58138-cb97-4641-b73f-aa3abb1aec45/volumes" Sep 30 16:29:17 crc kubenswrapper[4796]: I0930 16:29:17.028117 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerStarted","Data":"4ba7fa2bd8f1df00c3b5e72684dc40302b8e3f29ef82f6424b817662114ebbf1"} Sep 30 16:29:17 crc kubenswrapper[4796]: I0930 16:29:17.293239 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 16:29:17 crc kubenswrapper[4796]: W0930 16:29:17.294478 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf5d97fd_5337_41e8_9480_86b0342e9598.slice/crio-9dfc0813bddb70e4a139033bb633cf4551048ad4ad6a4d420b09774a0ffc3a3e WatchSource:0}: Error finding container 9dfc0813bddb70e4a139033bb633cf4551048ad4ad6a4d420b09774a0ffc3a3e: Status 404 returned error can't find the container with id 9dfc0813bddb70e4a139033bb633cf4551048ad4ad6a4d420b09774a0ffc3a3e Sep 30 16:29:17 crc kubenswrapper[4796]: I0930 16:29:17.992861 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:18 crc kubenswrapper[4796]: I0930 16:29:18.057580 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cf5d97fd-5337-41e8-9480-86b0342e9598","Type":"ContainerStarted","Data":"0f30ab9675ae2a75d6fcf83adcda8edbb6d9c99c8deda11ba6f6de5166ac4590"} Sep 30 16:29:18 crc kubenswrapper[4796]: I0930 16:29:18.057622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cf5d97fd-5337-41e8-9480-86b0342e9598","Type":"ContainerStarted","Data":"9dfc0813bddb70e4a139033bb633cf4551048ad4ad6a4d420b09774a0ffc3a3e"} Sep 30 16:29:18 crc kubenswrapper[4796]: I0930 16:29:18.060463 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerStarted","Data":"9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7"} Sep 30 16:29:18 crc kubenswrapper[4796]: I0930 16:29:18.060540 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerStarted","Data":"23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21"} Sep 30 16:29:19 crc kubenswrapper[4796]: I0930 16:29:19.070244 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cf5d97fd-5337-41e8-9480-86b0342e9598","Type":"ContainerStarted","Data":"bf7e36c8f4e4b53c0e9d79749c3d5fbe907f1f9e08a3b3d558247e31ac551aa8"} Sep 30 16:29:19 crc kubenswrapper[4796]: I0930 16:29:19.072609 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerStarted","Data":"729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0"} Sep 30 16:29:19 crc kubenswrapper[4796]: I0930 16:29:19.102748 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.102722323 podStartE2EDuration="3.102722323s" podCreationTimestamp="2025-09-30 16:29:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:19.087950911 +0000 UTC m=+1051.101229438" watchObservedRunningTime="2025-09-30 16:29:19.102722323 +0000 UTC m=+1051.116000860" Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.072417 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.073687 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85f459d7c9-l56gz" Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.085271 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerStarted","Data":"c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb"} Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.085544 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-central-agent" containerID="cri-o://23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21" gracePeriod=30 Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.085669 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="proxy-httpd" containerID="cri-o://c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb" gracePeriod=30 Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.085721 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="sg-core" containerID="cri-o://729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0" gracePeriod=30 Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.085766 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-notification-agent" containerID="cri-o://9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7" gracePeriod=30 Sep 30 16:29:20 crc kubenswrapper[4796]: I0930 16:29:20.171185 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.82128187 podStartE2EDuration="5.171164883s" podCreationTimestamp="2025-09-30 16:29:15 +0000 UTC" firstStartedPulling="2025-09-30 16:29:16.191272142 +0000 UTC m=+1048.204550669" lastFinishedPulling="2025-09-30 16:29:19.541155135 +0000 UTC m=+1051.554433682" observedRunningTime="2025-09-30 16:29:20.168194376 +0000 UTC m=+1052.181472903" watchObservedRunningTime="2025-09-30 16:29:20.171164883 +0000 UTC m=+1052.184443410" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.096400 4796 generic.go:334] "Generic (PLEG): container finished" podID="93081c0f-7233-4958-be78-a3cadc41d385" containerID="c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb" exitCode=0 Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.096631 4796 generic.go:334] "Generic (PLEG): container finished" podID="93081c0f-7233-4958-be78-a3cadc41d385" containerID="729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0" exitCode=2 Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.096638 4796 generic.go:334] "Generic (PLEG): container finished" podID="93081c0f-7233-4958-be78-a3cadc41d385" containerID="9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7" exitCode=0 Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.096546 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerDied","Data":"c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb"} Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.096719 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerDied","Data":"729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0"} Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.096735 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerDied","Data":"9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7"} Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.667533 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.848846 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958556 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-run-httpd\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958646 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-sg-core-conf-yaml\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958685 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xp6r\" (UniqueName: \"kubernetes.io/projected/93081c0f-7233-4958-be78-a3cadc41d385-kube-api-access-4xp6r\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958785 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-scripts\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958808 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-log-httpd\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958828 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-config-data\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.958872 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-combined-ca-bundle\") pod \"93081c0f-7233-4958-be78-a3cadc41d385\" (UID: \"93081c0f-7233-4958-be78-a3cadc41d385\") " Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.959081 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.959279 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.959303 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.965697 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93081c0f-7233-4958-be78-a3cadc41d385-kube-api-access-4xp6r" (OuterVolumeSpecName: "kube-api-access-4xp6r") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "kube-api-access-4xp6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.971152 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-scripts" (OuterVolumeSpecName: "scripts") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:21 crc kubenswrapper[4796]: I0930 16:29:21.993143 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.040753 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.061093 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.061119 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.061128 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xp6r\" (UniqueName: \"kubernetes.io/projected/93081c0f-7233-4958-be78-a3cadc41d385-kube-api-access-4xp6r\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.061138 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.061147 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93081c0f-7233-4958-be78-a3cadc41d385-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.067420 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-config-data" (OuterVolumeSpecName: "config-data") pod "93081c0f-7233-4958-be78-a3cadc41d385" (UID: "93081c0f-7233-4958-be78-a3cadc41d385"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.112000 4796 generic.go:334] "Generic (PLEG): container finished" podID="93081c0f-7233-4958-be78-a3cadc41d385" containerID="23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21" exitCode=0 Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.112053 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.112055 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerDied","Data":"23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21"} Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.112101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93081c0f-7233-4958-be78-a3cadc41d385","Type":"ContainerDied","Data":"4ba7fa2bd8f1df00c3b5e72684dc40302b8e3f29ef82f6424b817662114ebbf1"} Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.112118 4796 scope.go:117] "RemoveContainer" containerID="c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.160591 4796 scope.go:117] "RemoveContainer" containerID="729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.162724 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93081c0f-7233-4958-be78-a3cadc41d385-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.162772 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.173861 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190270 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.190635 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="sg-core" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190651 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="sg-core" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.190664 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-central-agent" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190671 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-central-agent" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.190684 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="proxy-httpd" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190690 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="proxy-httpd" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.190703 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-notification-agent" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190711 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-notification-agent" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190877 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-central-agent" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190894 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="sg-core" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190920 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="ceilometer-notification-agent" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.190930 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="93081c0f-7233-4958-be78-a3cadc41d385" containerName="proxy-httpd" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.192496 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.194607 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.194776 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.199646 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.201852 4796 scope.go:117] "RemoveContainer" containerID="9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.236907 4796 scope.go:117] "RemoveContainer" containerID="23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.258038 4796 scope.go:117] "RemoveContainer" containerID="c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.258439 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb\": container with ID starting with c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb not found: ID does not exist" containerID="c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.258472 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb"} err="failed to get container status \"c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb\": rpc error: code = NotFound desc = could not find container \"c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb\": container with ID starting with c4b25af259da6b0d3510b3583c6906106590f9dfae9ac76544fa8028dcce36cb not found: ID does not exist" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.258497 4796 scope.go:117] "RemoveContainer" containerID="729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.258779 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0\": container with ID starting with 729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0 not found: ID does not exist" containerID="729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.258815 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0"} err="failed to get container status \"729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0\": rpc error: code = NotFound desc = could not find container \"729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0\": container with ID starting with 729ef7d6b80e2c7c7e122205a7b88b53e96c58d27725aa44aaca2b657f3ab0f0 not found: ID does not exist" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.258839 4796 scope.go:117] "RemoveContainer" containerID="9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.259154 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7\": container with ID starting with 9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7 not found: ID does not exist" containerID="9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.259190 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7"} err="failed to get container status \"9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7\": rpc error: code = NotFound desc = could not find container \"9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7\": container with ID starting with 9a3ee471a2d25c2848ebd54ec7c20a4d00587297d1d5b68130358005c11ee4b7 not found: ID does not exist" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.259216 4796 scope.go:117] "RemoveContainer" containerID="23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21" Sep 30 16:29:22 crc kubenswrapper[4796]: E0930 16:29:22.259528 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21\": container with ID starting with 23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21 not found: ID does not exist" containerID="23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.259552 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21"} err="failed to get container status \"23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21\": rpc error: code = NotFound desc = could not find container \"23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21\": container with ID starting with 23f43639e68e81a8b9db4b096ac5849f5e146c5ebc776968d24302a2bec83e21 not found: ID does not exist" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366053 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-config-data\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366352 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-run-httpd\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366447 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366569 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366687 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-scripts\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366781 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-log-httpd\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.366879 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frkvw\" (UniqueName: \"kubernetes.io/projected/3b72a235-1ec9-4cb7-a671-38157f94a94d-kube-api-access-frkvw\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.469439 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-scripts\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.469753 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-log-httpd\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.469852 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frkvw\" (UniqueName: \"kubernetes.io/projected/3b72a235-1ec9-4cb7-a671-38157f94a94d-kube-api-access-frkvw\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.469954 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-config-data\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.470086 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-run-httpd\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.470167 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.470249 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.470537 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-log-httpd\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.470908 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-run-httpd\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.473518 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-scripts\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.475909 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.478259 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-config-data\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.480630 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.506339 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frkvw\" (UniqueName: \"kubernetes.io/projected/3b72a235-1ec9-4cb7-a671-38157f94a94d-kube-api-access-frkvw\") pod \"ceilometer-0\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.524416 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:22 crc kubenswrapper[4796]: I0930 16:29:22.753623 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93081c0f-7233-4958-be78-a3cadc41d385" path="/var/lib/kubelet/pods/93081c0f-7233-4958-be78-a3cadc41d385/volumes" Sep 30 16:29:23 crc kubenswrapper[4796]: W0930 16:29:23.033545 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b72a235_1ec9_4cb7_a671_38157f94a94d.slice/crio-3fe56cd0bd063358efd1005d53960737674acb8afdc99c5967548e01571bfc7d WatchSource:0}: Error finding container 3fe56cd0bd063358efd1005d53960737674acb8afdc99c5967548e01571bfc7d: Status 404 returned error can't find the container with id 3fe56cd0bd063358efd1005d53960737674acb8afdc99c5967548e01571bfc7d Sep 30 16:29:23 crc kubenswrapper[4796]: I0930 16:29:23.044127 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:23 crc kubenswrapper[4796]: I0930 16:29:23.124706 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerStarted","Data":"3fe56cd0bd063358efd1005d53960737674acb8afdc99c5967548e01571bfc7d"} Sep 30 16:29:24 crc kubenswrapper[4796]: I0930 16:29:24.137605 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerStarted","Data":"5c49a0f66e10eb19d3d33f9e3c347b6a83533fe38d06711dfd879da28e90d6ab"} Sep 30 16:29:24 crc kubenswrapper[4796]: I0930 16:29:24.691425 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:29:24 crc kubenswrapper[4796]: I0930 16:29:24.692132 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-httpd" containerID="cri-o://f3eb08f83f733b1498e53fc6064a5759d513c70b1eea1ce30ac6c495877461a7" gracePeriod=30 Sep 30 16:29:24 crc kubenswrapper[4796]: I0930 16:29:24.692107 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-log" containerID="cri-o://bc30eee3be60f47c3b1013f77c6ec08b3e60471c3aa80c489a063eb21a6470c2" gracePeriod=30 Sep 30 16:29:25 crc kubenswrapper[4796]: I0930 16:29:25.150549 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerStarted","Data":"ff35bcca93179fbc88089a1b50bdb3d0c7e8914f248e473a35a753d3d71d62e6"} Sep 30 16:29:25 crc kubenswrapper[4796]: I0930 16:29:25.153644 4796 generic.go:334] "Generic (PLEG): container finished" podID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerID="bc30eee3be60f47c3b1013f77c6ec08b3e60471c3aa80c489a063eb21a6470c2" exitCode=143 Sep 30 16:29:25 crc kubenswrapper[4796]: I0930 16:29:25.153679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14ffda24-fedb-45c8-af77-b9776ecc2e3c","Type":"ContainerDied","Data":"bc30eee3be60f47c3b1013f77c6ec08b3e60471c3aa80c489a063eb21a6470c2"} Sep 30 16:29:26 crc kubenswrapper[4796]: I0930 16:29:26.082250 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:29:26 crc kubenswrapper[4796]: I0930 16:29:26.853278 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:29:26 crc kubenswrapper[4796]: I0930 16:29:26.853848 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-log" containerID="cri-o://7b62ecc64657dc028863ee3ec7626c4bb282ca4474126411be13829493b65274" gracePeriod=30 Sep 30 16:29:26 crc kubenswrapper[4796]: I0930 16:29:26.854034 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-httpd" containerID="cri-o://697115e46b410831fc44124a45e13f95e4fc2393d6d3f2e6e096e82c6c89c7d1" gracePeriod=30 Sep 30 16:29:26 crc kubenswrapper[4796]: I0930 16:29:26.904745 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.183008 4796 generic.go:334] "Generic (PLEG): container finished" podID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerID="7b62ecc64657dc028863ee3ec7626c4bb282ca4474126411be13829493b65274" exitCode=143 Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.183029 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4","Type":"ContainerDied","Data":"7b62ecc64657dc028863ee3ec7626c4bb282ca4474126411be13829493b65274"} Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.185390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerStarted","Data":"9164b7ae2ac89c441efdcec484337765d4f5f7bbed529027b453b7699d2aa2c2"} Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.464233 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-494pn"] Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.465694 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.474008 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-494pn"] Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.558839 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jx95c"] Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.559967 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.568322 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jx95c"] Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.571109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j64x\" (UniqueName: \"kubernetes.io/projected/c70e4cae-e24d-418e-8f54-e377aa2708d3-kube-api-access-8j64x\") pod \"nova-api-db-create-494pn\" (UID: \"c70e4cae-e24d-418e-8f54-e377aa2708d3\") " pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.668925 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4jh9b"] Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.670042 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.672695 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbrjg\" (UniqueName: \"kubernetes.io/projected/df2a9cd6-ca55-412d-9862-c4a59b44fa54-kube-api-access-tbrjg\") pod \"nova-cell0-db-create-jx95c\" (UID: \"df2a9cd6-ca55-412d-9862-c4a59b44fa54\") " pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.672808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j64x\" (UniqueName: \"kubernetes.io/projected/c70e4cae-e24d-418e-8f54-e377aa2708d3-kube-api-access-8j64x\") pod \"nova-api-db-create-494pn\" (UID: \"c70e4cae-e24d-418e-8f54-e377aa2708d3\") " pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.679137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4jh9b"] Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.697658 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j64x\" (UniqueName: \"kubernetes.io/projected/c70e4cae-e24d-418e-8f54-e377aa2708d3-kube-api-access-8j64x\") pod \"nova-api-db-create-494pn\" (UID: \"c70e4cae-e24d-418e-8f54-e377aa2708d3\") " pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.774054 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbrjg\" (UniqueName: \"kubernetes.io/projected/df2a9cd6-ca55-412d-9862-c4a59b44fa54-kube-api-access-tbrjg\") pod \"nova-cell0-db-create-jx95c\" (UID: \"df2a9cd6-ca55-412d-9862-c4a59b44fa54\") " pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.774134 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpghh\" (UniqueName: \"kubernetes.io/projected/5c1cb023-494f-4318-8b3b-bc7d67977219-kube-api-access-xpghh\") pod \"nova-cell1-db-create-4jh9b\" (UID: \"5c1cb023-494f-4318-8b3b-bc7d67977219\") " pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.790590 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.796577 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbrjg\" (UniqueName: \"kubernetes.io/projected/df2a9cd6-ca55-412d-9862-c4a59b44fa54-kube-api-access-tbrjg\") pod \"nova-cell0-db-create-jx95c\" (UID: \"df2a9cd6-ca55-412d-9862-c4a59b44fa54\") " pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.874005 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.875220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpghh\" (UniqueName: \"kubernetes.io/projected/5c1cb023-494f-4318-8b3b-bc7d67977219-kube-api-access-xpghh\") pod \"nova-cell1-db-create-4jh9b\" (UID: \"5c1cb023-494f-4318-8b3b-bc7d67977219\") " pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.892546 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpghh\" (UniqueName: \"kubernetes.io/projected/5c1cb023-494f-4318-8b3b-bc7d67977219-kube-api-access-xpghh\") pod \"nova-cell1-db-create-4jh9b\" (UID: \"5c1cb023-494f-4318-8b3b-bc7d67977219\") " pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:27 crc kubenswrapper[4796]: I0930 16:29:27.991359 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.208441 4796 generic.go:334] "Generic (PLEG): container finished" podID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerID="f3eb08f83f733b1498e53fc6064a5759d513c70b1eea1ce30ac6c495877461a7" exitCode=0 Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.208480 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14ffda24-fedb-45c8-af77-b9776ecc2e3c","Type":"ContainerDied","Data":"f3eb08f83f733b1498e53fc6064a5759d513c70b1eea1ce30ac6c495877461a7"} Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.437198 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-494pn"] Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.530747 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.542798 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jx95c"] Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.683379 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4jh9b"] Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702406 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-scripts\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702471 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-httpd-run\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702507 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-config-data\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702548 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pbff\" (UniqueName: \"kubernetes.io/projected/14ffda24-fedb-45c8-af77-b9776ecc2e3c-kube-api-access-7pbff\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702626 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-public-tls-certs\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702648 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-logs\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702695 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-combined-ca-bundle\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.702721 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\" (UID: \"14ffda24-fedb-45c8-af77-b9776ecc2e3c\") " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.703854 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.704211 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-logs" (OuterVolumeSpecName: "logs") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.713790 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-scripts" (OuterVolumeSpecName: "scripts") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.723118 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.731157 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ffda24-fedb-45c8-af77-b9776ecc2e3c-kube-api-access-7pbff" (OuterVolumeSpecName: "kube-api-access-7pbff") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "kube-api-access-7pbff". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.775640 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-config-data" (OuterVolumeSpecName: "config-data") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.785147 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.805998 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806043 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806054 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806063 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806071 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806079 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pbff\" (UniqueName: \"kubernetes.io/projected/14ffda24-fedb-45c8-af77-b9776ecc2e3c-kube-api-access-7pbff\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806089 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14ffda24-fedb-45c8-af77-b9776ecc2e3c-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.806417 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "14ffda24-fedb-45c8-af77-b9776ecc2e3c" (UID: "14ffda24-fedb-45c8-af77-b9776ecc2e3c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.832902 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.908002 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14ffda24-fedb-45c8-af77-b9776ecc2e3c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.908030 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.909757 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-685c9578b5-5pzck" Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.934819 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.989100 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77489776dd-pcw55"] Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.989697 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77489776dd-pcw55" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-api" containerID="cri-o://0f90adcb820563eb04b8b73ad46757ae6f0643a598f15bb8e359da1f67832b51" gracePeriod=30 Sep 30 16:29:28 crc kubenswrapper[4796]: I0930 16:29:28.989865 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77489776dd-pcw55" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-httpd" containerID="cri-o://823ca9104ebb44966c6cfc1f353cb8d1ccff73e7bb8ffc7deef59bb6be84aa46" gracePeriod=30 Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.226780 4796 generic.go:334] "Generic (PLEG): container finished" podID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerID="823ca9104ebb44966c6cfc1f353cb8d1ccff73e7bb8ffc7deef59bb6be84aa46" exitCode=0 Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.226862 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77489776dd-pcw55" event={"ID":"3df0f7be-028e-4610-9787-14df0c3e5ea0","Type":"ContainerDied","Data":"823ca9104ebb44966c6cfc1f353cb8d1ccff73e7bb8ffc7deef59bb6be84aa46"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.228135 4796 generic.go:334] "Generic (PLEG): container finished" podID="c70e4cae-e24d-418e-8f54-e377aa2708d3" containerID="a611bccd55e2ffce8221c0381c93b8542ef5e03c5c7b9496d4ab9a9f2e2c453e" exitCode=0 Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.228178 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-494pn" event={"ID":"c70e4cae-e24d-418e-8f54-e377aa2708d3","Type":"ContainerDied","Data":"a611bccd55e2ffce8221c0381c93b8542ef5e03c5c7b9496d4ab9a9f2e2c453e"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.228203 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-494pn" event={"ID":"c70e4cae-e24d-418e-8f54-e377aa2708d3","Type":"ContainerStarted","Data":"072f5a4ffabc335a8f3981cedd883ba1f9fbf10284d87e4cb53d15155291b20f"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.231102 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14ffda24-fedb-45c8-af77-b9776ecc2e3c","Type":"ContainerDied","Data":"e9abb5ef61384095a83d2e2b096269b1a56decad59f9c2dbfb3026d1373a6cd0"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.231137 4796 scope.go:117] "RemoveContainer" containerID="f3eb08f83f733b1498e53fc6064a5759d513c70b1eea1ce30ac6c495877461a7" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.231205 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.235911 4796 generic.go:334] "Generic (PLEG): container finished" podID="5c1cb023-494f-4318-8b3b-bc7d67977219" containerID="95b0f331fcab6b992b9f0b3330f7d4889971201384740512d8c4870b3f20cde1" exitCode=0 Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.236019 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4jh9b" event={"ID":"5c1cb023-494f-4318-8b3b-bc7d67977219","Type":"ContainerDied","Data":"95b0f331fcab6b992b9f0b3330f7d4889971201384740512d8c4870b3f20cde1"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.236055 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4jh9b" event={"ID":"5c1cb023-494f-4318-8b3b-bc7d67977219","Type":"ContainerStarted","Data":"d4a849dc693e0a99d69161a0ab484b58606a33e25d47df0b418ea56a6a6aa991"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.239693 4796 generic.go:334] "Generic (PLEG): container finished" podID="df2a9cd6-ca55-412d-9862-c4a59b44fa54" containerID="86ef7588e5d2a520dc64bd5980f777c8575e3983691cfdee37470da0f88dd78a" exitCode=0 Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.239764 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jx95c" event={"ID":"df2a9cd6-ca55-412d-9862-c4a59b44fa54","Type":"ContainerDied","Data":"86ef7588e5d2a520dc64bd5980f777c8575e3983691cfdee37470da0f88dd78a"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.239795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jx95c" event={"ID":"df2a9cd6-ca55-412d-9862-c4a59b44fa54","Type":"ContainerStarted","Data":"8b6f5d13d0981b3beefbbf364a26b3093218ef9d6e146d236a39a9caa4523c47"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.244924 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerStarted","Data":"f815e939dbd769c0072cdbf2ccfe1a282f4b085553f0457800f6dbd978a01e5e"} Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.245661 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.274908 4796 scope.go:117] "RemoveContainer" containerID="bc30eee3be60f47c3b1013f77c6ec08b3e60471c3aa80c489a063eb21a6470c2" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.291239 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.009874023 podStartE2EDuration="7.291217785s" podCreationTimestamp="2025-09-30 16:29:22 +0000 UTC" firstStartedPulling="2025-09-30 16:29:23.035852516 +0000 UTC m=+1055.049131043" lastFinishedPulling="2025-09-30 16:29:28.317196278 +0000 UTC m=+1060.330474805" observedRunningTime="2025-09-30 16:29:29.291218845 +0000 UTC m=+1061.304497392" watchObservedRunningTime="2025-09-30 16:29:29.291217785 +0000 UTC m=+1061.304496312" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.313854 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.322435 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.347166 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:29:29 crc kubenswrapper[4796]: E0930 16:29:29.347540 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-httpd" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.347559 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-httpd" Sep 30 16:29:29 crc kubenswrapper[4796]: E0930 16:29:29.347582 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-log" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.347591 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-log" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.347794 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-log" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.347822 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" containerName="glance-httpd" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.348708 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.352580 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.352971 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.357356 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519291 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-logs\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519430 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-config-data\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519470 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519490 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-scripts\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519512 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519541 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfsw6\" (UniqueName: \"kubernetes.io/projected/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-kube-api-access-bfsw6\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.519595 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621362 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621478 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-logs\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621508 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621567 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-config-data\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621606 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621628 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-scripts\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621656 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621685 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfsw6\" (UniqueName: \"kubernetes.io/projected/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-kube-api-access-bfsw6\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.621857 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.622763 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-logs\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.622823 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.627793 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-scripts\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.628671 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.629097 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.630973 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-config-data\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.647694 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfsw6\" (UniqueName: \"kubernetes.io/projected/4cf6a891-18be-4f28-ad83-8a1f399d1cbd-kube-api-access-bfsw6\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.667504 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"4cf6a891-18be-4f28-ad83-8a1f399d1cbd\") " pod="openstack/glance-default-external-api-0" Sep 30 16:29:29 crc kubenswrapper[4796]: I0930 16:29:29.675742 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.310079 4796 generic.go:334] "Generic (PLEG): container finished" podID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerID="697115e46b410831fc44124a45e13f95e4fc2393d6d3f2e6e096e82c6c89c7d1" exitCode=0 Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.310677 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-central-agent" containerID="cri-o://5c49a0f66e10eb19d3d33f9e3c347b6a83533fe38d06711dfd879da28e90d6ab" gracePeriod=30 Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.310746 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4","Type":"ContainerDied","Data":"697115e46b410831fc44124a45e13f95e4fc2393d6d3f2e6e096e82c6c89c7d1"} Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.311221 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="proxy-httpd" containerID="cri-o://f815e939dbd769c0072cdbf2ccfe1a282f4b085553f0457800f6dbd978a01e5e" gracePeriod=30 Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.311311 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-notification-agent" containerID="cri-o://ff35bcca93179fbc88089a1b50bdb3d0c7e8914f248e473a35a753d3d71d62e6" gracePeriod=30 Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.311347 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="sg-core" containerID="cri-o://9164b7ae2ac89c441efdcec484337765d4f5f7bbed529027b453b7699d2aa2c2" gracePeriod=30 Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.353410 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.671334 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743127 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-scripts\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743220 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743252 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-config-data\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743274 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-httpd-run\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743336 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-logs\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743392 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-combined-ca-bundle\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743416 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-internal-tls-certs\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.743474 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q8dz\" (UniqueName: \"kubernetes.io/projected/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-kube-api-access-2q8dz\") pod \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\" (UID: \"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.744495 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.748682 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-logs" (OuterVolumeSpecName: "logs") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.751911 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-kube-api-access-2q8dz" (OuterVolumeSpecName: "kube-api-access-2q8dz") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "kube-api-access-2q8dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.771000 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ffda24-fedb-45c8-af77-b9776ecc2e3c" path="/var/lib/kubelet/pods/14ffda24-fedb-45c8-af77-b9776ecc2e3c/volumes" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.771647 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.772816 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-scripts" (OuterVolumeSpecName: "scripts") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.849739 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.849964 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q8dz\" (UniqueName: \"kubernetes.io/projected/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-kube-api-access-2q8dz\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.850167 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.850263 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.850347 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.857930 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.864660 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.888191 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.899162 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.931165 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-config-data" (OuterVolumeSpecName: "config-data") pod "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" (UID: "1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.951677 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpghh\" (UniqueName: \"kubernetes.io/projected/5c1cb023-494f-4318-8b3b-bc7d67977219-kube-api-access-xpghh\") pod \"5c1cb023-494f-4318-8b3b-bc7d67977219\" (UID: \"5c1cb023-494f-4318-8b3b-bc7d67977219\") " Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.952545 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.952575 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.952590 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.952604 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.955703 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c1cb023-494f-4318-8b3b-bc7d67977219-kube-api-access-xpghh" (OuterVolumeSpecName: "kube-api-access-xpghh") pod "5c1cb023-494f-4318-8b3b-bc7d67977219" (UID: "5c1cb023-494f-4318-8b3b-bc7d67977219"). InnerVolumeSpecName "kube-api-access-xpghh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:30 crc kubenswrapper[4796]: I0930 16:29:30.987518 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.008941 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.053837 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j64x\" (UniqueName: \"kubernetes.io/projected/c70e4cae-e24d-418e-8f54-e377aa2708d3-kube-api-access-8j64x\") pod \"c70e4cae-e24d-418e-8f54-e377aa2708d3\" (UID: \"c70e4cae-e24d-418e-8f54-e377aa2708d3\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.054075 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbrjg\" (UniqueName: \"kubernetes.io/projected/df2a9cd6-ca55-412d-9862-c4a59b44fa54-kube-api-access-tbrjg\") pod \"df2a9cd6-ca55-412d-9862-c4a59b44fa54\" (UID: \"df2a9cd6-ca55-412d-9862-c4a59b44fa54\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.054519 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpghh\" (UniqueName: \"kubernetes.io/projected/5c1cb023-494f-4318-8b3b-bc7d67977219-kube-api-access-xpghh\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.058802 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df2a9cd6-ca55-412d-9862-c4a59b44fa54-kube-api-access-tbrjg" (OuterVolumeSpecName: "kube-api-access-tbrjg") pod "df2a9cd6-ca55-412d-9862-c4a59b44fa54" (UID: "df2a9cd6-ca55-412d-9862-c4a59b44fa54"). InnerVolumeSpecName "kube-api-access-tbrjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.064707 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70e4cae-e24d-418e-8f54-e377aa2708d3-kube-api-access-8j64x" (OuterVolumeSpecName: "kube-api-access-8j64x") pod "c70e4cae-e24d-418e-8f54-e377aa2708d3" (UID: "c70e4cae-e24d-418e-8f54-e377aa2708d3"). InnerVolumeSpecName "kube-api-access-8j64x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.097309 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.097353 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.159053 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j64x\" (UniqueName: \"kubernetes.io/projected/c70e4cae-e24d-418e-8f54-e377aa2708d3-kube-api-access-8j64x\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.159084 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbrjg\" (UniqueName: \"kubernetes.io/projected/df2a9cd6-ca55-412d-9862-c4a59b44fa54-kube-api-access-tbrjg\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.320345 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4","Type":"ContainerDied","Data":"1f692b1bc147d4def228b6b87a6888e6edc6ba7dd14f912f8445de2a021625dd"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.320390 4796 scope.go:117] "RemoveContainer" containerID="697115e46b410831fc44124a45e13f95e4fc2393d6d3f2e6e096e82c6c89c7d1" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.320495 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.323912 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4jh9b" event={"ID":"5c1cb023-494f-4318-8b3b-bc7d67977219","Type":"ContainerDied","Data":"d4a849dc693e0a99d69161a0ab484b58606a33e25d47df0b418ea56a6a6aa991"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.323935 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4jh9b" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.323953 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4a849dc693e0a99d69161a0ab484b58606a33e25d47df0b418ea56a6a6aa991" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.326046 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jx95c" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.326053 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jx95c" event={"ID":"df2a9cd6-ca55-412d-9862-c4a59b44fa54","Type":"ContainerDied","Data":"8b6f5d13d0981b3beefbbf364a26b3093218ef9d6e146d236a39a9caa4523c47"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.326075 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6f5d13d0981b3beefbbf364a26b3093218ef9d6e146d236a39a9caa4523c47" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332674 4796 generic.go:334] "Generic (PLEG): container finished" podID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerID="f815e939dbd769c0072cdbf2ccfe1a282f4b085553f0457800f6dbd978a01e5e" exitCode=0 Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332706 4796 generic.go:334] "Generic (PLEG): container finished" podID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerID="9164b7ae2ac89c441efdcec484337765d4f5f7bbed529027b453b7699d2aa2c2" exitCode=2 Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332718 4796 generic.go:334] "Generic (PLEG): container finished" podID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerID="ff35bcca93179fbc88089a1b50bdb3d0c7e8914f248e473a35a753d3d71d62e6" exitCode=0 Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332728 4796 generic.go:334] "Generic (PLEG): container finished" podID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerID="5c49a0f66e10eb19d3d33f9e3c347b6a83533fe38d06711dfd879da28e90d6ab" exitCode=0 Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332770 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerDied","Data":"f815e939dbd769c0072cdbf2ccfe1a282f4b085553f0457800f6dbd978a01e5e"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332794 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerDied","Data":"9164b7ae2ac89c441efdcec484337765d4f5f7bbed529027b453b7699d2aa2c2"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332808 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerDied","Data":"ff35bcca93179fbc88089a1b50bdb3d0c7e8914f248e473a35a753d3d71d62e6"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332820 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerDied","Data":"5c49a0f66e10eb19d3d33f9e3c347b6a83533fe38d06711dfd879da28e90d6ab"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332834 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b72a235-1ec9-4cb7-a671-38157f94a94d","Type":"ContainerDied","Data":"3fe56cd0bd063358efd1005d53960737674acb8afdc99c5967548e01571bfc7d"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.332844 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fe56cd0bd063358efd1005d53960737674acb8afdc99c5967548e01571bfc7d" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.334193 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-494pn" event={"ID":"c70e4cae-e24d-418e-8f54-e377aa2708d3","Type":"ContainerDied","Data":"072f5a4ffabc335a8f3981cedd883ba1f9fbf10284d87e4cb53d15155291b20f"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.334219 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="072f5a4ffabc335a8f3981cedd883ba1f9fbf10284d87e4cb53d15155291b20f" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.334277 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-494pn" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.340337 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4cf6a891-18be-4f28-ad83-8a1f399d1cbd","Type":"ContainerStarted","Data":"bd964735394b485f140710484a9cac043c2f397000cdfcd189503fd35fd92bf0"} Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.344623 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.353545 4796 scope.go:117] "RemoveContainer" containerID="7b62ecc64657dc028863ee3ec7626c4bb282ca4474126411be13829493b65274" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.402086 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.415960 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430029 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430457 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df2a9cd6-ca55-412d-9862-c4a59b44fa54" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430474 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="df2a9cd6-ca55-412d-9862-c4a59b44fa54" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430488 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="proxy-httpd" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430495 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="proxy-httpd" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430512 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-notification-agent" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430518 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-notification-agent" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430532 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="sg-core" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430537 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="sg-core" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430552 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-log" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430557 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-log" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430571 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-httpd" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430577 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-httpd" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430592 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-central-agent" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430600 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-central-agent" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430615 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c1cb023-494f-4318-8b3b-bc7d67977219" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430621 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c1cb023-494f-4318-8b3b-bc7d67977219" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: E0930 16:29:31.430632 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70e4cae-e24d-418e-8f54-e377aa2708d3" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430638 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70e4cae-e24d-418e-8f54-e377aa2708d3" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430793 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="df2a9cd6-ca55-412d-9862-c4a59b44fa54" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430804 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-log" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430819 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="proxy-httpd" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430828 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-notification-agent" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430837 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="sg-core" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430845 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" containerName="glance-httpd" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430859 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70e4cae-e24d-418e-8f54-e377aa2708d3" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430869 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c1cb023-494f-4318-8b3b-bc7d67977219" containerName="mariadb-database-create" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.430879 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" containerName="ceilometer-central-agent" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.431905 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.436415 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.437162 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.458762 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463055 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-run-httpd\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463113 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-scripts\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463180 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-log-httpd\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463267 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-sg-core-conf-yaml\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463311 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-combined-ca-bundle\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463357 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frkvw\" (UniqueName: \"kubernetes.io/projected/3b72a235-1ec9-4cb7-a671-38157f94a94d-kube-api-access-frkvw\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.463466 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-config-data\") pod \"3b72a235-1ec9-4cb7-a671-38157f94a94d\" (UID: \"3b72a235-1ec9-4cb7-a671-38157f94a94d\") " Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.464400 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.464747 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.468361 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b72a235-1ec9-4cb7-a671-38157f94a94d-kube-api-access-frkvw" (OuterVolumeSpecName: "kube-api-access-frkvw") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "kube-api-access-frkvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.474049 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-scripts" (OuterVolumeSpecName: "scripts") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.506033 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565054 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565098 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbsnq\" (UniqueName: \"kubernetes.io/projected/63bfb928-2342-4432-be94-4f37611ae51e-kube-api-access-nbsnq\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565122 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565346 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565468 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63bfb928-2342-4432-be94-4f37611ae51e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565550 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63bfb928-2342-4432-be94-4f37611ae51e-logs\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565675 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565774 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565786 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frkvw\" (UniqueName: \"kubernetes.io/projected/3b72a235-1ec9-4cb7-a671-38157f94a94d-kube-api-access-frkvw\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565798 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565806 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.565833 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b72a235-1ec9-4cb7-a671-38157f94a94d-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.571770 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.611149 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-config-data" (OuterVolumeSpecName: "config-data") pod "3b72a235-1ec9-4cb7-a671-38157f94a94d" (UID: "3b72a235-1ec9-4cb7-a671-38157f94a94d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.667910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668004 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63bfb928-2342-4432-be94-4f37611ae51e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668043 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63bfb928-2342-4432-be94-4f37611ae51e-logs\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668063 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668104 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668154 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbsnq\" (UniqueName: \"kubernetes.io/projected/63bfb928-2342-4432-be94-4f37611ae51e-kube-api-access-nbsnq\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668196 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668267 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668280 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b72a235-1ec9-4cb7-a671-38157f94a94d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.668714 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.669303 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63bfb928-2342-4432-be94-4f37611ae51e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.675112 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.676572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.681253 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.682042 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63bfb928-2342-4432-be94-4f37611ae51e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.683304 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63bfb928-2342-4432-be94-4f37611ae51e-logs\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.684234 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbsnq\" (UniqueName: \"kubernetes.io/projected/63bfb928-2342-4432-be94-4f37611ae51e-kube-api-access-nbsnq\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.712156 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"63bfb928-2342-4432-be94-4f37611ae51e\") " pod="openstack/glance-default-internal-api-0" Sep 30 16:29:31 crc kubenswrapper[4796]: I0930 16:29:31.762210 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:32 crc kubenswrapper[4796]: W0930 16:29:32.298359 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63bfb928_2342_4432_be94_4f37611ae51e.slice/crio-66c14f2d2c2da4ef398d40664ba56ff6e1ad92f6628aff1f8855d415062d4204 WatchSource:0}: Error finding container 66c14f2d2c2da4ef398d40664ba56ff6e1ad92f6628aff1f8855d415062d4204: Status 404 returned error can't find the container with id 66c14f2d2c2da4ef398d40664ba56ff6e1ad92f6628aff1f8855d415062d4204 Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.298756 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.354697 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63bfb928-2342-4432-be94-4f37611ae51e","Type":"ContainerStarted","Data":"66c14f2d2c2da4ef398d40664ba56ff6e1ad92f6628aff1f8855d415062d4204"} Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.363174 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.363246 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4cf6a891-18be-4f28-ad83-8a1f399d1cbd","Type":"ContainerStarted","Data":"89c2e505cfa01ad085d334c0b06db3d39210d036d19a89ce7e0de9ae0e6c830a"} Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.364307 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4cf6a891-18be-4f28-ad83-8a1f399d1cbd","Type":"ContainerStarted","Data":"24bee6859b92d4ede440f8d2a02bc2bdda1c86451156f02e404340fd4332beed"} Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.380109 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.380090488 podStartE2EDuration="3.380090488s" podCreationTimestamp="2025-09-30 16:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:32.378235914 +0000 UTC m=+1064.391514441" watchObservedRunningTime="2025-09-30 16:29:32.380090488 +0000 UTC m=+1064.393369015" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.456512 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.462784 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.478428 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.480400 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.484808 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.484912 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.499637 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584414 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-run-httpd\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584479 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vllwg\" (UniqueName: \"kubernetes.io/projected/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-kube-api-access-vllwg\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584550 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584583 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-log-httpd\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-config-data\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.584644 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-scripts\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686066 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686129 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-log-httpd\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686158 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-config-data\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686178 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686211 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-scripts\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686259 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-run-httpd\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686302 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vllwg\" (UniqueName: \"kubernetes.io/projected/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-kube-api-access-vllwg\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686712 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-log-httpd\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.686901 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-run-httpd\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.692347 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.694939 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-config-data\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.695418 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.703349 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-scripts\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.703573 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vllwg\" (UniqueName: \"kubernetes.io/projected/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-kube-api-access-vllwg\") pod \"ceilometer-0\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " pod="openstack/ceilometer-0" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.744381 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4" path="/var/lib/kubelet/pods/1ad18c0a-d9a3-4d5a-bf78-06f8e95d61d4/volumes" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.745106 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b72a235-1ec9-4cb7-a671-38157f94a94d" path="/var/lib/kubelet/pods/3b72a235-1ec9-4cb7-a671-38157f94a94d/volumes" Sep 30 16:29:32 crc kubenswrapper[4796]: I0930 16:29:32.811585 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:33 crc kubenswrapper[4796]: I0930 16:29:33.277846 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:33 crc kubenswrapper[4796]: I0930 16:29:33.371257 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63bfb928-2342-4432-be94-4f37611ae51e","Type":"ContainerStarted","Data":"33bd01d5ae5233737a2960835ceb8aa0e207ab7cdc1b155f7d35ab9484499852"} Sep 30 16:29:33 crc kubenswrapper[4796]: I0930 16:29:33.375135 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerStarted","Data":"183d7febcc21a867207f5831d979eec3bbf811236546df192e95ea90f5faf2f8"} Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.407909 4796 generic.go:334] "Generic (PLEG): container finished" podID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerID="0f90adcb820563eb04b8b73ad46757ae6f0643a598f15bb8e359da1f67832b51" exitCode=0 Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.408365 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77489776dd-pcw55" event={"ID":"3df0f7be-028e-4610-9787-14df0c3e5ea0","Type":"ContainerDied","Data":"0f90adcb820563eb04b8b73ad46757ae6f0643a598f15bb8e359da1f67832b51"} Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.410239 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63bfb928-2342-4432-be94-4f37611ae51e","Type":"ContainerStarted","Data":"e32b43f55b4b295ac710abe73e453f158d6cf635c7fa1bd2197dba7df8a8a79f"} Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.418445 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerStarted","Data":"4f7719d455579dc9b871a0af8b753143cc7d17d8ccb47a3bd5646bc15a7c42f7"} Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.437912 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.4378884960000002 podStartE2EDuration="3.437888496s" podCreationTimestamp="2025-09-30 16:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:34.436804024 +0000 UTC m=+1066.450082581" watchObservedRunningTime="2025-09-30 16:29:34.437888496 +0000 UTC m=+1066.451167023" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.564260 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.640238 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-httpd-config\") pod \"3df0f7be-028e-4610-9787-14df0c3e5ea0\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.640343 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhx5x\" (UniqueName: \"kubernetes.io/projected/3df0f7be-028e-4610-9787-14df0c3e5ea0-kube-api-access-qhx5x\") pod \"3df0f7be-028e-4610-9787-14df0c3e5ea0\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.640369 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-config\") pod \"3df0f7be-028e-4610-9787-14df0c3e5ea0\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.640390 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-combined-ca-bundle\") pod \"3df0f7be-028e-4610-9787-14df0c3e5ea0\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.640528 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-ovndb-tls-certs\") pod \"3df0f7be-028e-4610-9787-14df0c3e5ea0\" (UID: \"3df0f7be-028e-4610-9787-14df0c3e5ea0\") " Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.648636 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df0f7be-028e-4610-9787-14df0c3e5ea0-kube-api-access-qhx5x" (OuterVolumeSpecName: "kube-api-access-qhx5x") pod "3df0f7be-028e-4610-9787-14df0c3e5ea0" (UID: "3df0f7be-028e-4610-9787-14df0c3e5ea0"). InnerVolumeSpecName "kube-api-access-qhx5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.649551 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3df0f7be-028e-4610-9787-14df0c3e5ea0" (UID: "3df0f7be-028e-4610-9787-14df0c3e5ea0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.702083 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3df0f7be-028e-4610-9787-14df0c3e5ea0" (UID: "3df0f7be-028e-4610-9787-14df0c3e5ea0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.706768 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-config" (OuterVolumeSpecName: "config") pod "3df0f7be-028e-4610-9787-14df0c3e5ea0" (UID: "3df0f7be-028e-4610-9787-14df0c3e5ea0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.730551 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "3df0f7be-028e-4610-9787-14df0c3e5ea0" (UID: "3df0f7be-028e-4610-9787-14df0c3e5ea0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.742913 4796 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.743109 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.743197 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhx5x\" (UniqueName: \"kubernetes.io/projected/3df0f7be-028e-4610-9787-14df0c3e5ea0-kube-api-access-qhx5x\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.743280 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:34 crc kubenswrapper[4796]: I0930 16:29:34.743386 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0f7be-028e-4610-9787-14df0c3e5ea0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.074518 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149098 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-combined-ca-bundle\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149164 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-etc-machine-id\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149187 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjjfv\" (UniqueName: \"kubernetes.io/projected/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-kube-api-access-zjjfv\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149280 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data-custom\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149329 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149382 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-logs\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.149433 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-scripts\") pod \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\" (UID: \"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8\") " Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.152671 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.153022 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-logs" (OuterVolumeSpecName: "logs") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.157551 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-scripts" (OuterVolumeSpecName: "scripts") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.160200 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.162185 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-kube-api-access-zjjfv" (OuterVolumeSpecName: "kube-api-access-zjjfv") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "kube-api-access-zjjfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.189844 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.237009 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data" (OuterVolumeSpecName: "config-data") pod "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" (UID: "8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252105 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252141 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252150 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252158 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252170 4796 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252180 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjjfv\" (UniqueName: \"kubernetes.io/projected/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-kube-api-access-zjjfv\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.252190 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.442118 4796 generic.go:334] "Generic (PLEG): container finished" podID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerID="e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6" exitCode=137 Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.442211 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8","Type":"ContainerDied","Data":"e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6"} Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.442246 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8","Type":"ContainerDied","Data":"c0311f8e903d7f7727143319f269b655f376c0455d30ff3712ad51bea0e5eea5"} Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.442265 4796 scope.go:117] "RemoveContainer" containerID="e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.442291 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.447134 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77489776dd-pcw55" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.447615 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77489776dd-pcw55" event={"ID":"3df0f7be-028e-4610-9787-14df0c3e5ea0","Type":"ContainerDied","Data":"ede8dd1c81966e279b43e72afb68ff51827104393f7d620c46534e09d4a267e6"} Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.475655 4796 scope.go:117] "RemoveContainer" containerID="8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.517677 4796 scope.go:117] "RemoveContainer" containerID="e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6" Sep 30 16:29:35 crc kubenswrapper[4796]: E0930 16:29:35.520086 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6\": container with ID starting with e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6 not found: ID does not exist" containerID="e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.520162 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6"} err="failed to get container status \"e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6\": rpc error: code = NotFound desc = could not find container \"e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6\": container with ID starting with e68f571b192c7b12908bd1e587b8af68fd75b0888d214f377777cdef9191afd6 not found: ID does not exist" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.520199 4796 scope.go:117] "RemoveContainer" containerID="8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06" Sep 30 16:29:35 crc kubenswrapper[4796]: E0930 16:29:35.524002 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06\": container with ID starting with 8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06 not found: ID does not exist" containerID="8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.524072 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06"} err="failed to get container status \"8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06\": rpc error: code = NotFound desc = could not find container \"8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06\": container with ID starting with 8d44be047b15acfe215da40e41e17865986f16e060a902dd7b48f66736cdbe06 not found: ID does not exist" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.524110 4796 scope.go:117] "RemoveContainer" containerID="823ca9104ebb44966c6cfc1f353cb8d1ccff73e7bb8ffc7deef59bb6be84aa46" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.526487 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77489776dd-pcw55"] Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.556875 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-77489776dd-pcw55"] Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.563600 4796 scope.go:117] "RemoveContainer" containerID="0f90adcb820563eb04b8b73ad46757ae6f0643a598f15bb8e359da1f67832b51" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.568492 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.576783 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.591053 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:35 crc kubenswrapper[4796]: E0930 16:29:35.591601 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.591631 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api" Sep 30 16:29:35 crc kubenswrapper[4796]: E0930 16:29:35.591659 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-httpd" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.591670 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-httpd" Sep 30 16:29:35 crc kubenswrapper[4796]: E0930 16:29:35.591687 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api-log" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.591695 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api-log" Sep 30 16:29:35 crc kubenswrapper[4796]: E0930 16:29:35.591715 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-api" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.591724 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-api" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.591951 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-httpd" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.592006 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api-log" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.592023 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" containerName="neutron-api" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.592040 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" containerName="cinder-api" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.593764 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.599476 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.601941 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.602156 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.602290 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.661872 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-config-data\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662232 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e2552e-914c-48e2-99c1-06b6ef87aefe-logs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662284 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77kp6\" (UniqueName: \"kubernetes.io/projected/89e2552e-914c-48e2-99c1-06b6ef87aefe-kube-api-access-77kp6\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662326 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/89e2552e-914c-48e2-99c1-06b6ef87aefe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662357 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-config-data-custom\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662377 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662567 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-scripts\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662627 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.662681 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.764901 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-scripts\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.764967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765032 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765100 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-config-data\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765139 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e2552e-914c-48e2-99c1-06b6ef87aefe-logs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765195 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77kp6\" (UniqueName: \"kubernetes.io/projected/89e2552e-914c-48e2-99c1-06b6ef87aefe-kube-api-access-77kp6\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765268 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/89e2552e-914c-48e2-99c1-06b6ef87aefe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765291 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-config-data-custom\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765322 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765547 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e2552e-914c-48e2-99c1-06b6ef87aefe-logs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.765610 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/89e2552e-914c-48e2-99c1-06b6ef87aefe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.770770 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-scripts\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.771237 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-config-data-custom\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.772392 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.772411 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-config-data\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.775042 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.776329 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e2552e-914c-48e2-99c1-06b6ef87aefe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.788613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77kp6\" (UniqueName: \"kubernetes.io/projected/89e2552e-914c-48e2-99c1-06b6ef87aefe-kube-api-access-77kp6\") pod \"cinder-api-0\" (UID: \"89e2552e-914c-48e2-99c1-06b6ef87aefe\") " pod="openstack/cinder-api-0" Sep 30 16:29:35 crc kubenswrapper[4796]: I0930 16:29:35.920500 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 16:29:36 crc kubenswrapper[4796]: I0930 16:29:36.401419 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 16:29:36 crc kubenswrapper[4796]: W0930 16:29:36.403496 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89e2552e_914c_48e2_99c1_06b6ef87aefe.slice/crio-e922c73574c3f87dedbf170f6971c84a349bc0841996f0047b150fac31c6569c WatchSource:0}: Error finding container e922c73574c3f87dedbf170f6971c84a349bc0841996f0047b150fac31c6569c: Status 404 returned error can't find the container with id e922c73574c3f87dedbf170f6971c84a349bc0841996f0047b150fac31c6569c Sep 30 16:29:36 crc kubenswrapper[4796]: I0930 16:29:36.484910 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerStarted","Data":"ccd938973993a4177a0b14e1a3b45f28fb61a0b32bd6205649a312b33ecd73ac"} Sep 30 16:29:36 crc kubenswrapper[4796]: I0930 16:29:36.484958 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerStarted","Data":"d7c900b179c515b84e1ea896fdf4a076fb9ae1cd1c1fa3a159cb9864f9dadec6"} Sep 30 16:29:36 crc kubenswrapper[4796]: I0930 16:29:36.494365 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"89e2552e-914c-48e2-99c1-06b6ef87aefe","Type":"ContainerStarted","Data":"e922c73574c3f87dedbf170f6971c84a349bc0841996f0047b150fac31c6569c"} Sep 30 16:29:36 crc kubenswrapper[4796]: I0930 16:29:36.744304 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df0f7be-028e-4610-9787-14df0c3e5ea0" path="/var/lib/kubelet/pods/3df0f7be-028e-4610-9787-14df0c3e5ea0/volumes" Sep 30 16:29:36 crc kubenswrapper[4796]: I0930 16:29:36.745217 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8" path="/var/lib/kubelet/pods/8ba6cafc-d5fb-4ac3-b29f-69ffe11a3aa8/volumes" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.518361 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"89e2552e-914c-48e2-99c1-06b6ef87aefe","Type":"ContainerStarted","Data":"1f9ac465f69e045d6d5e4289fe325b66370efab4dd98e43a00a791fe43a2cf06"} Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.642493 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6979-account-create-nlhts"] Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.644214 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.647156 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.652796 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6979-account-create-nlhts"] Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.795139 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5721-account-create-6k8xp"] Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.796259 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.798155 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.808830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv62j\" (UniqueName: \"kubernetes.io/projected/087eaf83-481b-4b24-a3e8-3a674de51028-kube-api-access-fv62j\") pod \"nova-api-6979-account-create-nlhts\" (UID: \"087eaf83-481b-4b24-a3e8-3a674de51028\") " pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.838215 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5721-account-create-6k8xp"] Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.911207 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv62j\" (UniqueName: \"kubernetes.io/projected/087eaf83-481b-4b24-a3e8-3a674de51028-kube-api-access-fv62j\") pod \"nova-api-6979-account-create-nlhts\" (UID: \"087eaf83-481b-4b24-a3e8-3a674de51028\") " pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.911347 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhhd\" (UniqueName: \"kubernetes.io/projected/16e579e1-4e1c-4b03-9a22-af7399019ac4-kube-api-access-dkhhd\") pod \"nova-cell0-5721-account-create-6k8xp\" (UID: \"16e579e1-4e1c-4b03-9a22-af7399019ac4\") " pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.932515 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv62j\" (UniqueName: \"kubernetes.io/projected/087eaf83-481b-4b24-a3e8-3a674de51028-kube-api-access-fv62j\") pod \"nova-api-6979-account-create-nlhts\" (UID: \"087eaf83-481b-4b24-a3e8-3a674de51028\") " pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.972751 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:37 crc kubenswrapper[4796]: I0930 16:29:37.998076 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1fcc-account-create-ngpsd"] Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.000045 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.005551 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.011277 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1fcc-account-create-ngpsd"] Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.017310 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhhd\" (UniqueName: \"kubernetes.io/projected/16e579e1-4e1c-4b03-9a22-af7399019ac4-kube-api-access-dkhhd\") pod \"nova-cell0-5721-account-create-6k8xp\" (UID: \"16e579e1-4e1c-4b03-9a22-af7399019ac4\") " pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.037771 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhhd\" (UniqueName: \"kubernetes.io/projected/16e579e1-4e1c-4b03-9a22-af7399019ac4-kube-api-access-dkhhd\") pod \"nova-cell0-5721-account-create-6k8xp\" (UID: \"16e579e1-4e1c-4b03-9a22-af7399019ac4\") " pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.119505 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8n9k\" (UniqueName: \"kubernetes.io/projected/7fee951a-80a3-4086-a18a-7121e4402a07-kube-api-access-x8n9k\") pod \"nova-cell1-1fcc-account-create-ngpsd\" (UID: \"7fee951a-80a3-4086-a18a-7121e4402a07\") " pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.184713 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.221483 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8n9k\" (UniqueName: \"kubernetes.io/projected/7fee951a-80a3-4086-a18a-7121e4402a07-kube-api-access-x8n9k\") pod \"nova-cell1-1fcc-account-create-ngpsd\" (UID: \"7fee951a-80a3-4086-a18a-7121e4402a07\") " pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.242941 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8n9k\" (UniqueName: \"kubernetes.io/projected/7fee951a-80a3-4086-a18a-7121e4402a07-kube-api-access-x8n9k\") pod \"nova-cell1-1fcc-account-create-ngpsd\" (UID: \"7fee951a-80a3-4086-a18a-7121e4402a07\") " pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.338613 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:38 crc kubenswrapper[4796]: W0930 16:29:38.495606 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod087eaf83_481b_4b24_a3e8_3a674de51028.slice/crio-a4b479d2cbfd54588752912246e0330972637a4d91e98846db4bf570be45857c WatchSource:0}: Error finding container a4b479d2cbfd54588752912246e0330972637a4d91e98846db4bf570be45857c: Status 404 returned error can't find the container with id a4b479d2cbfd54588752912246e0330972637a4d91e98846db4bf570be45857c Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.496495 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6979-account-create-nlhts"] Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.532123 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerStarted","Data":"8037976ee42397f5959c634ed72bee1207e9625f855f69c6a3f28fca889adb6b"} Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.534419 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6979-account-create-nlhts" event={"ID":"087eaf83-481b-4b24-a3e8-3a674de51028","Type":"ContainerStarted","Data":"a4b479d2cbfd54588752912246e0330972637a4d91e98846db4bf570be45857c"} Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.535946 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"89e2552e-914c-48e2-99c1-06b6ef87aefe","Type":"ContainerStarted","Data":"9a329ece95e9db86de3f256954afab792e0776ce8a090dca8ebe877484023329"} Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.536085 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.566712 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.145693923 podStartE2EDuration="6.566688865s" podCreationTimestamp="2025-09-30 16:29:32 +0000 UTC" firstStartedPulling="2025-09-30 16:29:33.290426592 +0000 UTC m=+1065.303705119" lastFinishedPulling="2025-09-30 16:29:37.711421534 +0000 UTC m=+1069.724700061" observedRunningTime="2025-09-30 16:29:38.55763347 +0000 UTC m=+1070.570911997" watchObservedRunningTime="2025-09-30 16:29:38.566688865 +0000 UTC m=+1070.579967412" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.583699 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.583675792 podStartE2EDuration="3.583675792s" podCreationTimestamp="2025-09-30 16:29:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:29:38.57574771 +0000 UTC m=+1070.589026287" watchObservedRunningTime="2025-09-30 16:29:38.583675792 +0000 UTC m=+1070.596954329" Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.625994 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5721-account-create-6k8xp"] Sep 30 16:29:38 crc kubenswrapper[4796]: W0930 16:29:38.628575 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16e579e1_4e1c_4b03_9a22_af7399019ac4.slice/crio-773723a8fb717cbcfcb20797f1ba4af8cccb30554206cd5e6556c132758e1ea2 WatchSource:0}: Error finding container 773723a8fb717cbcfcb20797f1ba4af8cccb30554206cd5e6556c132758e1ea2: Status 404 returned error can't find the container with id 773723a8fb717cbcfcb20797f1ba4af8cccb30554206cd5e6556c132758e1ea2 Sep 30 16:29:38 crc kubenswrapper[4796]: I0930 16:29:38.765197 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1fcc-account-create-ngpsd"] Sep 30 16:29:38 crc kubenswrapper[4796]: W0930 16:29:38.794459 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fee951a_80a3_4086_a18a_7121e4402a07.slice/crio-f34421f502df9919da31f97a8be620039ab2a89c7f524de58315f46b2d41d92f WatchSource:0}: Error finding container f34421f502df9919da31f97a8be620039ab2a89c7f524de58315f46b2d41d92f: Status 404 returned error can't find the container with id f34421f502df9919da31f97a8be620039ab2a89c7f524de58315f46b2d41d92f Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.547359 4796 generic.go:334] "Generic (PLEG): container finished" podID="16e579e1-4e1c-4b03-9a22-af7399019ac4" containerID="b87b361f05d2973a3fdd3eedb2e30570cfc502408e856dc1877050a44cfd2eba" exitCode=0 Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.547401 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5721-account-create-6k8xp" event={"ID":"16e579e1-4e1c-4b03-9a22-af7399019ac4","Type":"ContainerDied","Data":"b87b361f05d2973a3fdd3eedb2e30570cfc502408e856dc1877050a44cfd2eba"} Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.547448 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5721-account-create-6k8xp" event={"ID":"16e579e1-4e1c-4b03-9a22-af7399019ac4","Type":"ContainerStarted","Data":"773723a8fb717cbcfcb20797f1ba4af8cccb30554206cd5e6556c132758e1ea2"} Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.552055 4796 generic.go:334] "Generic (PLEG): container finished" podID="087eaf83-481b-4b24-a3e8-3a674de51028" containerID="b92956c52884a939123cf26ef0ee6bee755252ed7edb05ee5551d1fa800820df" exitCode=0 Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.552100 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6979-account-create-nlhts" event={"ID":"087eaf83-481b-4b24-a3e8-3a674de51028","Type":"ContainerDied","Data":"b92956c52884a939123cf26ef0ee6bee755252ed7edb05ee5551d1fa800820df"} Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.554217 4796 generic.go:334] "Generic (PLEG): container finished" podID="7fee951a-80a3-4086-a18a-7121e4402a07" containerID="98c13e2cf727be9d106a4804a2e659d109f6f410e018070c6d6f97def1af8a85" exitCode=0 Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.554303 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" event={"ID":"7fee951a-80a3-4086-a18a-7121e4402a07","Type":"ContainerDied","Data":"98c13e2cf727be9d106a4804a2e659d109f6f410e018070c6d6f97def1af8a85"} Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.554342 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" event={"ID":"7fee951a-80a3-4086-a18a-7121e4402a07","Type":"ContainerStarted","Data":"f34421f502df9919da31f97a8be620039ab2a89c7f524de58315f46b2d41d92f"} Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.554695 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.677055 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.677356 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.715083 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 16:29:39 crc kubenswrapper[4796]: I0930 16:29:39.754436 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 16:29:40 crc kubenswrapper[4796]: I0930 16:29:40.568910 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 16:29:40 crc kubenswrapper[4796]: I0930 16:29:40.569151 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 16:29:40 crc kubenswrapper[4796]: I0930 16:29:40.963617 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.073265 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8n9k\" (UniqueName: \"kubernetes.io/projected/7fee951a-80a3-4086-a18a-7121e4402a07-kube-api-access-x8n9k\") pod \"7fee951a-80a3-4086-a18a-7121e4402a07\" (UID: \"7fee951a-80a3-4086-a18a-7121e4402a07\") " Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.080282 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fee951a-80a3-4086-a18a-7121e4402a07-kube-api-access-x8n9k" (OuterVolumeSpecName: "kube-api-access-x8n9k") pod "7fee951a-80a3-4086-a18a-7121e4402a07" (UID: "7fee951a-80a3-4086-a18a-7121e4402a07"). InnerVolumeSpecName "kube-api-access-x8n9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.155104 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.165212 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.175472 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8n9k\" (UniqueName: \"kubernetes.io/projected/7fee951a-80a3-4086-a18a-7121e4402a07-kube-api-access-x8n9k\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.276850 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkhhd\" (UniqueName: \"kubernetes.io/projected/16e579e1-4e1c-4b03-9a22-af7399019ac4-kube-api-access-dkhhd\") pod \"16e579e1-4e1c-4b03-9a22-af7399019ac4\" (UID: \"16e579e1-4e1c-4b03-9a22-af7399019ac4\") " Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.277154 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv62j\" (UniqueName: \"kubernetes.io/projected/087eaf83-481b-4b24-a3e8-3a674de51028-kube-api-access-fv62j\") pod \"087eaf83-481b-4b24-a3e8-3a674de51028\" (UID: \"087eaf83-481b-4b24-a3e8-3a674de51028\") " Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.280283 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16e579e1-4e1c-4b03-9a22-af7399019ac4-kube-api-access-dkhhd" (OuterVolumeSpecName: "kube-api-access-dkhhd") pod "16e579e1-4e1c-4b03-9a22-af7399019ac4" (UID: "16e579e1-4e1c-4b03-9a22-af7399019ac4"). InnerVolumeSpecName "kube-api-access-dkhhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.280481 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087eaf83-481b-4b24-a3e8-3a674de51028-kube-api-access-fv62j" (OuterVolumeSpecName: "kube-api-access-fv62j") pod "087eaf83-481b-4b24-a3e8-3a674de51028" (UID: "087eaf83-481b-4b24-a3e8-3a674de51028"). InnerVolumeSpecName "kube-api-access-fv62j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.380175 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkhhd\" (UniqueName: \"kubernetes.io/projected/16e579e1-4e1c-4b03-9a22-af7399019ac4-kube-api-access-dkhhd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.380209 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv62j\" (UniqueName: \"kubernetes.io/projected/087eaf83-481b-4b24-a3e8-3a674de51028-kube-api-access-fv62j\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.580847 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6979-account-create-nlhts" event={"ID":"087eaf83-481b-4b24-a3e8-3a674de51028","Type":"ContainerDied","Data":"a4b479d2cbfd54588752912246e0330972637a4d91e98846db4bf570be45857c"} Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.580894 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b479d2cbfd54588752912246e0330972637a4d91e98846db4bf570be45857c" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.580902 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6979-account-create-nlhts" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.583110 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" event={"ID":"7fee951a-80a3-4086-a18a-7121e4402a07","Type":"ContainerDied","Data":"f34421f502df9919da31f97a8be620039ab2a89c7f524de58315f46b2d41d92f"} Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.583303 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f34421f502df9919da31f97a8be620039ab2a89c7f524de58315f46b2d41d92f" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.583126 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1fcc-account-create-ngpsd" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.585166 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5721-account-create-6k8xp" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.585755 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5721-account-create-6k8xp" event={"ID":"16e579e1-4e1c-4b03-9a22-af7399019ac4","Type":"ContainerDied","Data":"773723a8fb717cbcfcb20797f1ba4af8cccb30554206cd5e6556c132758e1ea2"} Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.585799 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="773723a8fb717cbcfcb20797f1ba4af8cccb30554206cd5e6556c132758e1ea2" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.763128 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.763372 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.797113 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:41 crc kubenswrapper[4796]: I0930 16:29:41.828443 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.043497 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.044031 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-notification-agent" containerID="cri-o://d7c900b179c515b84e1ea896fdf4a076fb9ae1cd1c1fa3a159cb9864f9dadec6" gracePeriod=30 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.044031 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="sg-core" containerID="cri-o://ccd938973993a4177a0b14e1a3b45f28fb61a0b32bd6205649a312b33ecd73ac" gracePeriod=30 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.044247 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-central-agent" containerID="cri-o://4f7719d455579dc9b871a0af8b753143cc7d17d8ccb47a3bd5646bc15a7c42f7" gracePeriod=30 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.044038 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="proxy-httpd" containerID="cri-o://8037976ee42397f5959c634ed72bee1207e9625f855f69c6a3f28fca889adb6b" gracePeriod=30 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.403797 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.404102 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599174 4796 generic.go:334] "Generic (PLEG): container finished" podID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerID="8037976ee42397f5959c634ed72bee1207e9625f855f69c6a3f28fca889adb6b" exitCode=0 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599448 4796 generic.go:334] "Generic (PLEG): container finished" podID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerID="ccd938973993a4177a0b14e1a3b45f28fb61a0b32bd6205649a312b33ecd73ac" exitCode=2 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599447 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerDied","Data":"8037976ee42397f5959c634ed72bee1207e9625f855f69c6a3f28fca889adb6b"} Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599513 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerDied","Data":"ccd938973993a4177a0b14e1a3b45f28fb61a0b32bd6205649a312b33ecd73ac"} Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerDied","Data":"d7c900b179c515b84e1ea896fdf4a076fb9ae1cd1c1fa3a159cb9864f9dadec6"} Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599464 4796 generic.go:334] "Generic (PLEG): container finished" podID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerID="d7c900b179c515b84e1ea896fdf4a076fb9ae1cd1c1fa3a159cb9864f9dadec6" exitCode=0 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.599548 4796 generic.go:334] "Generic (PLEG): container finished" podID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerID="4f7719d455579dc9b871a0af8b753143cc7d17d8ccb47a3bd5646bc15a7c42f7" exitCode=0 Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.600720 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerDied","Data":"4f7719d455579dc9b871a0af8b753143cc7d17d8ccb47a3bd5646bc15a7c42f7"} Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.600756 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.600769 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.803578 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908314 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-scripts\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908388 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-sg-core-conf-yaml\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908487 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-run-httpd\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908536 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vllwg\" (UniqueName: \"kubernetes.io/projected/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-kube-api-access-vllwg\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908560 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-log-httpd\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908596 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-config-data\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908613 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-combined-ca-bundle\") pod \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\" (UID: \"c51a9fe7-d19e-4f30-9201-94f55cef6e5d\") " Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908893 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.908928 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.909182 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.909201 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.930870 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-scripts" (OuterVolumeSpecName: "scripts") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.933123 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-kube-api-access-vllwg" (OuterVolumeSpecName: "kube-api-access-vllwg") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "kube-api-access-vllwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:29:42 crc kubenswrapper[4796]: I0930 16:29:42.937670 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.011632 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.011662 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.011672 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vllwg\" (UniqueName: \"kubernetes.io/projected/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-kube-api-access-vllwg\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.026832 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-config-data" (OuterVolumeSpecName: "config-data") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.078612 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c51a9fe7-d19e-4f30-9201-94f55cef6e5d" (UID: "c51a9fe7-d19e-4f30-9201-94f55cef6e5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098475 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fvd2d"] Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098839 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="sg-core" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098849 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="sg-core" Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098858 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="proxy-httpd" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098863 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="proxy-httpd" Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098881 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fee951a-80a3-4086-a18a-7121e4402a07" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098889 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fee951a-80a3-4086-a18a-7121e4402a07" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098904 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087eaf83-481b-4b24-a3e8-3a674de51028" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098910 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="087eaf83-481b-4b24-a3e8-3a674de51028" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098928 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-notification-agent" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098935 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-notification-agent" Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098950 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-central-agent" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.098957 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-central-agent" Sep 30 16:29:43 crc kubenswrapper[4796]: E0930 16:29:43.098967 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16e579e1-4e1c-4b03-9a22-af7399019ac4" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099054 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="16e579e1-4e1c-4b03-9a22-af7399019ac4" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099246 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fee951a-80a3-4086-a18a-7121e4402a07" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099269 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-notification-agent" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099282 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="sg-core" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099291 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="proxy-httpd" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099302 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" containerName="ceilometer-central-agent" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099312 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="16e579e1-4e1c-4b03-9a22-af7399019ac4" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099323 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="087eaf83-481b-4b24-a3e8-3a674de51028" containerName="mariadb-account-create" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.099881 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.117248 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ffcj6" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.117561 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.117888 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.121605 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fvd2d"] Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.123412 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.123456 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51a9fe7-d19e-4f30-9201-94f55cef6e5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.225546 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-scripts\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.225681 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.225958 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqzzh\" (UniqueName: \"kubernetes.io/projected/1e4ce38d-a600-48e0-8178-676b1bbd17c1-kube-api-access-lqzzh\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.226142 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-config-data\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.328067 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqzzh\" (UniqueName: \"kubernetes.io/projected/1e4ce38d-a600-48e0-8178-676b1bbd17c1-kube-api-access-lqzzh\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.328132 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-config-data\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.328169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-scripts\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.328202 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.332672 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.332691 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-config-data\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.332876 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-scripts\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.345511 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqzzh\" (UniqueName: \"kubernetes.io/projected/1e4ce38d-a600-48e0-8178-676b1bbd17c1-kube-api-access-lqzzh\") pod \"nova-cell0-conductor-db-sync-fvd2d\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.441517 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.630419 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c51a9fe7-d19e-4f30-9201-94f55cef6e5d","Type":"ContainerDied","Data":"183d7febcc21a867207f5831d979eec3bbf811236546df192e95ea90f5faf2f8"} Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.630729 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.630735 4796 scope.go:117] "RemoveContainer" containerID="8037976ee42397f5959c634ed72bee1207e9625f855f69c6a3f28fca889adb6b" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.672888 4796 scope.go:117] "RemoveContainer" containerID="ccd938973993a4177a0b14e1a3b45f28fb61a0b32bd6205649a312b33ecd73ac" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.676302 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.682486 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.700480 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.702493 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.708072 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.708234 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.721768 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.725229 4796 scope.go:117] "RemoveContainer" containerID="d7c900b179c515b84e1ea896fdf4a076fb9ae1cd1c1fa3a159cb9864f9dadec6" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.772136 4796 scope.go:117] "RemoveContainer" containerID="4f7719d455579dc9b871a0af8b753143cc7d17d8ccb47a3bd5646bc15a7c42f7" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.840479 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.840530 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-run-httpd\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.840550 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-config-data\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.840578 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29h8q\" (UniqueName: \"kubernetes.io/projected/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-kube-api-access-29h8q\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.840823 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-scripts\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.840951 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-log-httpd\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.841500 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.906133 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fvd2d"] Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.942836 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.942887 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-run-httpd\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.942906 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-config-data\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.942933 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29h8q\" (UniqueName: \"kubernetes.io/projected/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-kube-api-access-29h8q\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.942998 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-scripts\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.943033 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-log-httpd\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.943098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.943788 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-run-httpd\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.947376 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-log-httpd\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.952109 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-scripts\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.954541 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.955506 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.955861 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-config-data\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:43 crc kubenswrapper[4796]: I0930 16:29:43.963674 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29h8q\" (UniqueName: \"kubernetes.io/projected/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-kube-api-access-29h8q\") pod \"ceilometer-0\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " pod="openstack/ceilometer-0" Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.026476 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.521606 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.528965 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.645200 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerStarted","Data":"e0e5b0c8f4d0100ec1807f2d6fac250b17ea0282f8e9fa777c55b60247e7a519"} Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.647821 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.648685 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" event={"ID":"1e4ce38d-a600-48e0-8178-676b1bbd17c1","Type":"ContainerStarted","Data":"82c55bee13e285316b82ac43088b4692d0e0e51e1ddb992b31941185b11767c3"} Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.746922 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51a9fe7-d19e-4f30-9201-94f55cef6e5d" path="/var/lib/kubelet/pods/c51a9fe7-d19e-4f30-9201-94f55cef6e5d/volumes" Sep 30 16:29:44 crc kubenswrapper[4796]: I0930 16:29:44.809946 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 16:29:45 crc kubenswrapper[4796]: I0930 16:29:45.660512 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerStarted","Data":"720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56"} Sep 30 16:29:46 crc kubenswrapper[4796]: I0930 16:29:46.673196 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerStarted","Data":"7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e"} Sep 30 16:29:47 crc kubenswrapper[4796]: I0930 16:29:47.684964 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerStarted","Data":"c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3"} Sep 30 16:29:47 crc kubenswrapper[4796]: I0930 16:29:47.862536 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 16:29:51 crc kubenswrapper[4796]: I0930 16:29:51.725853 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerStarted","Data":"8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0"} Sep 30 16:29:51 crc kubenswrapper[4796]: I0930 16:29:51.726483 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:29:51 crc kubenswrapper[4796]: I0930 16:29:51.727556 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" event={"ID":"1e4ce38d-a600-48e0-8178-676b1bbd17c1","Type":"ContainerStarted","Data":"c1eb20b2801e7239d54a8f0baee6f4a3f6b073d266e3fddc78bb29a987925897"} Sep 30 16:29:51 crc kubenswrapper[4796]: I0930 16:29:51.773687 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.187442209 podStartE2EDuration="8.773660221s" podCreationTimestamp="2025-09-30 16:29:43 +0000 UTC" firstStartedPulling="2025-09-30 16:29:44.556233665 +0000 UTC m=+1076.569512192" lastFinishedPulling="2025-09-30 16:29:51.142451637 +0000 UTC m=+1083.155730204" observedRunningTime="2025-09-30 16:29:51.760009711 +0000 UTC m=+1083.773288238" watchObservedRunningTime="2025-09-30 16:29:51.773660221 +0000 UTC m=+1083.786938788" Sep 30 16:29:51 crc kubenswrapper[4796]: I0930 16:29:51.783641 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" podStartSLOduration=1.5432330840000001 podStartE2EDuration="8.783621102s" podCreationTimestamp="2025-09-30 16:29:43 +0000 UTC" firstStartedPulling="2025-09-30 16:29:43.918481989 +0000 UTC m=+1075.931760516" lastFinishedPulling="2025-09-30 16:29:51.158869957 +0000 UTC m=+1083.172148534" observedRunningTime="2025-09-30 16:29:51.774376302 +0000 UTC m=+1083.787654859" watchObservedRunningTime="2025-09-30 16:29:51.783621102 +0000 UTC m=+1083.796899659" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.142733 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x"] Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.145363 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.148182 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.148309 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.153356 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x"] Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.257969 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64df679-26d4-4957-b7da-a922e6fecae7-config-volume\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.258091 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsttr\" (UniqueName: \"kubernetes.io/projected/a64df679-26d4-4957-b7da-a922e6fecae7-kube-api-access-lsttr\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.258188 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64df679-26d4-4957-b7da-a922e6fecae7-secret-volume\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.359430 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64df679-26d4-4957-b7da-a922e6fecae7-config-volume\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.359486 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsttr\" (UniqueName: \"kubernetes.io/projected/a64df679-26d4-4957-b7da-a922e6fecae7-kube-api-access-lsttr\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.359573 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64df679-26d4-4957-b7da-a922e6fecae7-secret-volume\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.360396 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64df679-26d4-4957-b7da-a922e6fecae7-config-volume\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.370025 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64df679-26d4-4957-b7da-a922e6fecae7-secret-volume\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.381226 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsttr\" (UniqueName: \"kubernetes.io/projected/a64df679-26d4-4957-b7da-a922e6fecae7-kube-api-access-lsttr\") pod \"collect-profiles-29320830-b5f8x\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.467259 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:00 crc kubenswrapper[4796]: W0930 16:30:00.944409 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda64df679_26d4_4957_b7da_a922e6fecae7.slice/crio-472090154abb8a851f674a5d80eca097577d670b38cad97307c2c6a9310678d7 WatchSource:0}: Error finding container 472090154abb8a851f674a5d80eca097577d670b38cad97307c2c6a9310678d7: Status 404 returned error can't find the container with id 472090154abb8a851f674a5d80eca097577d670b38cad97307c2c6a9310678d7 Sep 30 16:30:00 crc kubenswrapper[4796]: I0930 16:30:00.946709 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x"] Sep 30 16:30:01 crc kubenswrapper[4796]: I0930 16:30:01.097382 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:30:01 crc kubenswrapper[4796]: I0930 16:30:01.097814 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:30:01 crc kubenswrapper[4796]: I0930 16:30:01.856958 4796 generic.go:334] "Generic (PLEG): container finished" podID="a64df679-26d4-4957-b7da-a922e6fecae7" containerID="14223aa6f8d1e508171e72c57257bfded8bc71030ccfb4005775e7c9b0fcc815" exitCode=0 Sep 30 16:30:01 crc kubenswrapper[4796]: I0930 16:30:01.857032 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" event={"ID":"a64df679-26d4-4957-b7da-a922e6fecae7","Type":"ContainerDied","Data":"14223aa6f8d1e508171e72c57257bfded8bc71030ccfb4005775e7c9b0fcc815"} Sep 30 16:30:01 crc kubenswrapper[4796]: I0930 16:30:01.857287 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" event={"ID":"a64df679-26d4-4957-b7da-a922e6fecae7","Type":"ContainerStarted","Data":"472090154abb8a851f674a5d80eca097577d670b38cad97307c2c6a9310678d7"} Sep 30 16:30:02 crc kubenswrapper[4796]: I0930 16:30:02.870632 4796 generic.go:334] "Generic (PLEG): container finished" podID="1e4ce38d-a600-48e0-8178-676b1bbd17c1" containerID="c1eb20b2801e7239d54a8f0baee6f4a3f6b073d266e3fddc78bb29a987925897" exitCode=0 Sep 30 16:30:02 crc kubenswrapper[4796]: I0930 16:30:02.870700 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" event={"ID":"1e4ce38d-a600-48e0-8178-676b1bbd17c1","Type":"ContainerDied","Data":"c1eb20b2801e7239d54a8f0baee6f4a3f6b073d266e3fddc78bb29a987925897"} Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.216810 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.328804 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64df679-26d4-4957-b7da-a922e6fecae7-config-volume\") pod \"a64df679-26d4-4957-b7da-a922e6fecae7\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.328939 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64df679-26d4-4957-b7da-a922e6fecae7-secret-volume\") pod \"a64df679-26d4-4957-b7da-a922e6fecae7\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.329042 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsttr\" (UniqueName: \"kubernetes.io/projected/a64df679-26d4-4957-b7da-a922e6fecae7-kube-api-access-lsttr\") pod \"a64df679-26d4-4957-b7da-a922e6fecae7\" (UID: \"a64df679-26d4-4957-b7da-a922e6fecae7\") " Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.329700 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a64df679-26d4-4957-b7da-a922e6fecae7-config-volume" (OuterVolumeSpecName: "config-volume") pod "a64df679-26d4-4957-b7da-a922e6fecae7" (UID: "a64df679-26d4-4957-b7da-a922e6fecae7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.334271 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64df679-26d4-4957-b7da-a922e6fecae7-kube-api-access-lsttr" (OuterVolumeSpecName: "kube-api-access-lsttr") pod "a64df679-26d4-4957-b7da-a922e6fecae7" (UID: "a64df679-26d4-4957-b7da-a922e6fecae7"). InnerVolumeSpecName "kube-api-access-lsttr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.334953 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64df679-26d4-4957-b7da-a922e6fecae7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a64df679-26d4-4957-b7da-a922e6fecae7" (UID: "a64df679-26d4-4957-b7da-a922e6fecae7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.431037 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64df679-26d4-4957-b7da-a922e6fecae7-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.431082 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64df679-26d4-4957-b7da-a922e6fecae7-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.431092 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsttr\" (UniqueName: \"kubernetes.io/projected/a64df679-26d4-4957-b7da-a922e6fecae7-kube-api-access-lsttr\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.889600 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" event={"ID":"a64df679-26d4-4957-b7da-a922e6fecae7","Type":"ContainerDied","Data":"472090154abb8a851f674a5d80eca097577d670b38cad97307c2c6a9310678d7"} Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.890096 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="472090154abb8a851f674a5d80eca097577d670b38cad97307c2c6a9310678d7" Sep 30 16:30:03 crc kubenswrapper[4796]: I0930 16:30:03.890318 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320830-b5f8x" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.232121 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.350119 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-config-data\") pod \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.350424 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqzzh\" (UniqueName: \"kubernetes.io/projected/1e4ce38d-a600-48e0-8178-676b1bbd17c1-kube-api-access-lqzzh\") pod \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.350795 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-scripts\") pod \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.350970 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-combined-ca-bundle\") pod \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\" (UID: \"1e4ce38d-a600-48e0-8178-676b1bbd17c1\") " Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.355350 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e4ce38d-a600-48e0-8178-676b1bbd17c1-kube-api-access-lqzzh" (OuterVolumeSpecName: "kube-api-access-lqzzh") pod "1e4ce38d-a600-48e0-8178-676b1bbd17c1" (UID: "1e4ce38d-a600-48e0-8178-676b1bbd17c1"). InnerVolumeSpecName "kube-api-access-lqzzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.356097 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-scripts" (OuterVolumeSpecName: "scripts") pod "1e4ce38d-a600-48e0-8178-676b1bbd17c1" (UID: "1e4ce38d-a600-48e0-8178-676b1bbd17c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.383604 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-config-data" (OuterVolumeSpecName: "config-data") pod "1e4ce38d-a600-48e0-8178-676b1bbd17c1" (UID: "1e4ce38d-a600-48e0-8178-676b1bbd17c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.389265 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e4ce38d-a600-48e0-8178-676b1bbd17c1" (UID: "1e4ce38d-a600-48e0-8178-676b1bbd17c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.453608 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.453966 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.454030 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4ce38d-a600-48e0-8178-676b1bbd17c1-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.454047 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqzzh\" (UniqueName: \"kubernetes.io/projected/1e4ce38d-a600-48e0-8178-676b1bbd17c1-kube-api-access-lqzzh\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.901073 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" event={"ID":"1e4ce38d-a600-48e0-8178-676b1bbd17c1","Type":"ContainerDied","Data":"82c55bee13e285316b82ac43088b4692d0e0e51e1ddb992b31941185b11767c3"} Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.901124 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82c55bee13e285316b82ac43088b4692d0e0e51e1ddb992b31941185b11767c3" Sep 30 16:30:04 crc kubenswrapper[4796]: I0930 16:30:04.901136 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fvd2d" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.006315 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 16:30:05 crc kubenswrapper[4796]: E0930 16:30:05.006754 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e4ce38d-a600-48e0-8178-676b1bbd17c1" containerName="nova-cell0-conductor-db-sync" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.006774 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e4ce38d-a600-48e0-8178-676b1bbd17c1" containerName="nova-cell0-conductor-db-sync" Sep 30 16:30:05 crc kubenswrapper[4796]: E0930 16:30:05.006806 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64df679-26d4-4957-b7da-a922e6fecae7" containerName="collect-profiles" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.006815 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64df679-26d4-4957-b7da-a922e6fecae7" containerName="collect-profiles" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.007056 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64df679-26d4-4957-b7da-a922e6fecae7" containerName="collect-profiles" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.007079 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e4ce38d-a600-48e0-8178-676b1bbd17c1" containerName="nova-cell0-conductor-db-sync" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.007761 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.016536 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ffcj6" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.016587 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.017472 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.168256 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875eff01-1202-4914-8f8b-172f661da26a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.168353 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hjvz\" (UniqueName: \"kubernetes.io/projected/875eff01-1202-4914-8f8b-172f661da26a-kube-api-access-9hjvz\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.168417 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875eff01-1202-4914-8f8b-172f661da26a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.270120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875eff01-1202-4914-8f8b-172f661da26a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.270313 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875eff01-1202-4914-8f8b-172f661da26a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.270365 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hjvz\" (UniqueName: \"kubernetes.io/projected/875eff01-1202-4914-8f8b-172f661da26a-kube-api-access-9hjvz\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.274657 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875eff01-1202-4914-8f8b-172f661da26a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.274709 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/875eff01-1202-4914-8f8b-172f661da26a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.295111 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hjvz\" (UniqueName: \"kubernetes.io/projected/875eff01-1202-4914-8f8b-172f661da26a-kube-api-access-9hjvz\") pod \"nova-cell0-conductor-0\" (UID: \"875eff01-1202-4914-8f8b-172f661da26a\") " pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.340307 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.866594 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 16:30:05 crc kubenswrapper[4796]: W0930 16:30:05.877922 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod875eff01_1202_4914_8f8b_172f661da26a.slice/crio-64e5205971356a0f03b1a5ec52e0c4df8a512af20483d03e91a816f2b61e2a10 WatchSource:0}: Error finding container 64e5205971356a0f03b1a5ec52e0c4df8a512af20483d03e91a816f2b61e2a10: Status 404 returned error can't find the container with id 64e5205971356a0f03b1a5ec52e0c4df8a512af20483d03e91a816f2b61e2a10 Sep 30 16:30:05 crc kubenswrapper[4796]: I0930 16:30:05.911647 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"875eff01-1202-4914-8f8b-172f661da26a","Type":"ContainerStarted","Data":"64e5205971356a0f03b1a5ec52e0c4df8a512af20483d03e91a816f2b61e2a10"} Sep 30 16:30:06 crc kubenswrapper[4796]: I0930 16:30:06.930821 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"875eff01-1202-4914-8f8b-172f661da26a","Type":"ContainerStarted","Data":"009b3bfaa342424074b067366fe4c131be6fe9dc61336eb67e0a71b054098234"} Sep 30 16:30:06 crc kubenswrapper[4796]: I0930 16:30:06.931235 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:06 crc kubenswrapper[4796]: I0930 16:30:06.959609 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.959589194 podStartE2EDuration="2.959589194s" podCreationTimestamp="2025-09-30 16:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:06.949372745 +0000 UTC m=+1098.962651272" watchObservedRunningTime="2025-09-30 16:30:06.959589194 +0000 UTC m=+1098.972867721" Sep 30 16:30:14 crc kubenswrapper[4796]: I0930 16:30:14.033698 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 16:30:15 crc kubenswrapper[4796]: I0930 16:30:15.389173 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 30 16:30:15 crc kubenswrapper[4796]: I0930 16:30:15.949152 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-wlmls"] Sep 30 16:30:15 crc kubenswrapper[4796]: I0930 16:30:15.950364 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:15 crc kubenswrapper[4796]: I0930 16:30:15.952332 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 30 16:30:15 crc kubenswrapper[4796]: I0930 16:30:15.953417 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 30 16:30:15 crc kubenswrapper[4796]: I0930 16:30:15.963861 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlmls"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.030169 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.030256 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-scripts\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.030388 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhwzd\" (UniqueName: \"kubernetes.io/projected/f10a7487-db94-441a-856d-dff951ed56cf-kube-api-access-zhwzd\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.030471 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-config-data\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.079211 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.080663 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.082924 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.095849 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.133083 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhwzd\" (UniqueName: \"kubernetes.io/projected/f10a7487-db94-441a-856d-dff951ed56cf-kube-api-access-zhwzd\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.133171 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-config-data\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.133201 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.133237 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-scripts\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.141278 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-config-data\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.142924 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.150424 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.153970 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-scripts\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.154055 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.160309 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.172838 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhwzd\" (UniqueName: \"kubernetes.io/projected/f10a7487-db94-441a-856d-dff951ed56cf-kube-api-access-zhwzd\") pod \"nova-cell0-cell-mapping-wlmls\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.192109 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235430 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235481 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9c6\" (UniqueName: \"kubernetes.io/projected/5357c107-43c4-46a5-b422-e1caca9bf032-kube-api-access-4w9c6\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235508 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5357c107-43c4-46a5-b422-e1caca9bf032-logs\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235571 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljjsc\" (UniqueName: \"kubernetes.io/projected/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-kube-api-access-ljjsc\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235603 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-config-data\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235628 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-config-data\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.235677 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.274719 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.288466 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.289880 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.296479 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.335814 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336787 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336811 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9c6\" (UniqueName: \"kubernetes.io/projected/5357c107-43c4-46a5-b422-e1caca9bf032-kube-api-access-4w9c6\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336832 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5357c107-43c4-46a5-b422-e1caca9bf032-logs\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336874 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljjsc\" (UniqueName: \"kubernetes.io/projected/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-kube-api-access-ljjsc\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336895 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-config-data\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336913 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-config-data\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.336944 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.337549 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5357c107-43c4-46a5-b422-e1caca9bf032-logs\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.344936 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.346090 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-config-data\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.347109 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.359584 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-config-data\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.378403 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljjsc\" (UniqueName: \"kubernetes.io/projected/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-kube-api-access-ljjsc\") pod \"nova-scheduler-0\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.407252 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9c6\" (UniqueName: \"kubernetes.io/projected/5357c107-43c4-46a5-b422-e1caca9bf032-kube-api-access-4w9c6\") pod \"nova-api-0\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.428431 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.429687 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.431398 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.440619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2cfd\" (UniqueName: \"kubernetes.io/projected/919fb585-87e9-4502-90ee-a225c034f707-kube-api-access-l2cfd\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.440704 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919fb585-87e9-4502-90ee-a225c034f707-logs\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.440989 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.444312 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-config-data\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.455341 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.485070 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hn7wq"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.487475 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.511189 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hn7wq"] Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.545694 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.545931 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-config-data\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.545998 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2cfd\" (UniqueName: \"kubernetes.io/projected/919fb585-87e9-4502-90ee-a225c034f707-kube-api-access-l2cfd\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.546025 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mmw2\" (UniqueName: \"kubernetes.io/projected/720f2613-b6eb-43df-95a2-d05f980bb0c0-kube-api-access-5mmw2\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.546053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919fb585-87e9-4502-90ee-a225c034f707-logs\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.546074 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.546096 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.546663 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919fb585-87e9-4502-90ee-a225c034f707-logs\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.553072 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-config-data\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.554160 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.555662 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.570646 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2cfd\" (UniqueName: \"kubernetes.io/projected/919fb585-87e9-4502-90ee-a225c034f707-kube-api-access-l2cfd\") pod \"nova-metadata-0\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-config\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647154 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647178 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647198 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9m69\" (UniqueName: \"kubernetes.io/projected/c3e1c88b-b3be-443d-a971-7433e3d868e7-kube-api-access-p9m69\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647259 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647330 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647368 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647398 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mmw2\" (UniqueName: \"kubernetes.io/projected/720f2613-b6eb-43df-95a2-d05f980bb0c0-kube-api-access-5mmw2\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.647422 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.667972 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.668134 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.704968 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mmw2\" (UniqueName: \"kubernetes.io/projected/720f2613-b6eb-43df-95a2-d05f980bb0c0-kube-api-access-5mmw2\") pod \"nova-cell1-novncproxy-0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.712534 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.748810 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.748879 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.748911 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.748960 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-config\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.748998 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9m69\" (UniqueName: \"kubernetes.io/projected/c3e1c88b-b3be-443d-a971-7433e3d868e7-kube-api-access-p9m69\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.749058 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.750296 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.752362 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.753451 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.754214 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.754353 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-config\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.772622 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9m69\" (UniqueName: \"kubernetes.io/projected/c3e1c88b-b3be-443d-a971-7433e3d868e7-kube-api-access-p9m69\") pod \"dnsmasq-dns-845d6d6f59-hn7wq\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.792380 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.838568 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:16 crc kubenswrapper[4796]: I0930 16:30:16.845530 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.001834 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jpscz"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.025398 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlmls"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.025510 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.031354 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.033899 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jpscz"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.043216 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.164131 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-scripts\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.164447 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56rkt\" (UniqueName: \"kubernetes.io/projected/800b5cc2-8806-4298-802a-7bef00b36a4c-kube-api-access-56rkt\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.164490 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.164679 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-config-data\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.186731 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.266768 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-scripts\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.266823 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56rkt\" (UniqueName: \"kubernetes.io/projected/800b5cc2-8806-4298-802a-7bef00b36a4c-kube-api-access-56rkt\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.266850 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.266957 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-config-data\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.270424 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-config-data\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.281248 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-scripts\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.282927 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.295050 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56rkt\" (UniqueName: \"kubernetes.io/projected/800b5cc2-8806-4298-802a-7bef00b36a4c-kube-api-access-56rkt\") pod \"nova-cell1-conductor-db-sync-jpscz\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.343806 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.395598 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.469489 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.756954 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hn7wq"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.776656 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:17 crc kubenswrapper[4796]: I0930 16:30:17.965557 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jpscz"] Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.058767 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"720f2613-b6eb-43df-95a2-d05f980bb0c0","Type":"ContainerStarted","Data":"9ae8c12c6553751e748202aad4d487b2137c98a19705db11ae3115ac15f872e1"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.061449 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlmls" event={"ID":"f10a7487-db94-441a-856d-dff951ed56cf","Type":"ContainerStarted","Data":"a57aa13db2e426dfd044edd4989d291ba10b52a8d7e36eb6324e3600e091bc27"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.061476 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlmls" event={"ID":"f10a7487-db94-441a-856d-dff951ed56cf","Type":"ContainerStarted","Data":"7c13497c2b2bdb3ae6766feb417de9af623f0b310fc0889a9cdab32d5466ed70"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.065115 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jpscz" event={"ID":"800b5cc2-8806-4298-802a-7bef00b36a4c","Type":"ContainerStarted","Data":"e8eb7a4f52c537b8bc32cbbb975b7149117a7634f3a59708eca8ee9fa8b08b01"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.066737 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3","Type":"ContainerStarted","Data":"579249f88007949d974e24497896e264b1ed3054caf2118befebe1cadd1745cf"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.068699 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" event={"ID":"c3e1c88b-b3be-443d-a971-7433e3d868e7","Type":"ContainerStarted","Data":"151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.068733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" event={"ID":"c3e1c88b-b3be-443d-a971-7433e3d868e7","Type":"ContainerStarted","Data":"34ceb3283afbf24ee1049b216c0e82d3711dfc94528092a7150e0b2f177bf621"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.077857 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"919fb585-87e9-4502-90ee-a225c034f707","Type":"ContainerStarted","Data":"c0ed9d3b940dcd989d30f8f30351ce3d763e95593cda973d9e03da47c6ef751c"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.086276 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5357c107-43c4-46a5-b422-e1caca9bf032","Type":"ContainerStarted","Data":"e42fec612e50a69c843ba7696eb57a8cda5d126bcd23edb8c50fd67534f90b44"} Sep 30 16:30:18 crc kubenswrapper[4796]: I0930 16:30:18.086743 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-wlmls" podStartSLOduration=3.086726528 podStartE2EDuration="3.086726528s" podCreationTimestamp="2025-09-30 16:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:18.082052411 +0000 UTC m=+1110.095330938" watchObservedRunningTime="2025-09-30 16:30:18.086726528 +0000 UTC m=+1110.100005055" Sep 30 16:30:19 crc kubenswrapper[4796]: I0930 16:30:19.102255 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jpscz" event={"ID":"800b5cc2-8806-4298-802a-7bef00b36a4c","Type":"ContainerStarted","Data":"d1e335a5c446dfe36da0709c7daadfd01ef39f4982921030c7fc75692148aa8e"} Sep 30 16:30:19 crc kubenswrapper[4796]: I0930 16:30:19.108773 4796 generic.go:334] "Generic (PLEG): container finished" podID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerID="151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9" exitCode=0 Sep 30 16:30:19 crc kubenswrapper[4796]: I0930 16:30:19.109151 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" event={"ID":"c3e1c88b-b3be-443d-a971-7433e3d868e7","Type":"ContainerDied","Data":"151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9"} Sep 30 16:30:19 crc kubenswrapper[4796]: I0930 16:30:19.128248 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-jpscz" podStartSLOduration=3.128228131 podStartE2EDuration="3.128228131s" podCreationTimestamp="2025-09-30 16:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:19.12375662 +0000 UTC m=+1111.137035157" watchObservedRunningTime="2025-09-30 16:30:19.128228131 +0000 UTC m=+1111.141506658" Sep 30 16:30:19 crc kubenswrapper[4796]: I0930 16:30:19.969310 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:19 crc kubenswrapper[4796]: I0930 16:30:19.981835 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.149241 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"720f2613-b6eb-43df-95a2-d05f980bb0c0","Type":"ContainerStarted","Data":"33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.150180 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="720f2613-b6eb-43df-95a2-d05f980bb0c0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7" gracePeriod=30 Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.154939 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3","Type":"ContainerStarted","Data":"a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.167022 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" event={"ID":"c3e1c88b-b3be-443d-a971-7433e3d868e7","Type":"ContainerStarted","Data":"e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.167092 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.170693 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"919fb585-87e9-4502-90ee-a225c034f707","Type":"ContainerStarted","Data":"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.170741 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"919fb585-87e9-4502-90ee-a225c034f707","Type":"ContainerStarted","Data":"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.175138 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-metadata" containerID="cri-o://7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942" gracePeriod=30 Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.172702 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-log" containerID="cri-o://0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e" gracePeriod=30 Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.181941 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.16580347 podStartE2EDuration="6.181919445s" podCreationTimestamp="2025-09-30 16:30:16 +0000 UTC" firstStartedPulling="2025-09-30 16:30:17.782773042 +0000 UTC m=+1109.796051569" lastFinishedPulling="2025-09-30 16:30:20.798889017 +0000 UTC m=+1112.812167544" observedRunningTime="2025-09-30 16:30:22.17563058 +0000 UTC m=+1114.188909117" watchObservedRunningTime="2025-09-30 16:30:22.181919445 +0000 UTC m=+1114.195197972" Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.188898 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5357c107-43c4-46a5-b422-e1caca9bf032","Type":"ContainerStarted","Data":"c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.188942 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5357c107-43c4-46a5-b422-e1caca9bf032","Type":"ContainerStarted","Data":"4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819"} Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.199254 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" podStartSLOduration=6.199235151 podStartE2EDuration="6.199235151s" podCreationTimestamp="2025-09-30 16:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:22.195043319 +0000 UTC m=+1114.208321846" watchObservedRunningTime="2025-09-30 16:30:22.199235151 +0000 UTC m=+1114.212513678" Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.228629 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.929978507 podStartE2EDuration="6.228601811s" podCreationTimestamp="2025-09-30 16:30:16 +0000 UTC" firstStartedPulling="2025-09-30 16:30:17.501782578 +0000 UTC m=+1109.515061105" lastFinishedPulling="2025-09-30 16:30:20.800405882 +0000 UTC m=+1112.813684409" observedRunningTime="2025-09-30 16:30:22.21831685 +0000 UTC m=+1114.231595367" watchObservedRunningTime="2025-09-30 16:30:22.228601811 +0000 UTC m=+1114.241880368" Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.242594 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.699218835 podStartE2EDuration="6.24257982s" podCreationTimestamp="2025-09-30 16:30:16 +0000 UTC" firstStartedPulling="2025-09-30 16:30:17.257023876 +0000 UTC m=+1109.270302403" lastFinishedPulling="2025-09-30 16:30:20.800384861 +0000 UTC m=+1112.813663388" observedRunningTime="2025-09-30 16:30:22.23505048 +0000 UTC m=+1114.248329027" watchObservedRunningTime="2025-09-30 16:30:22.24257982 +0000 UTC m=+1114.255858337" Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.258296 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.838038187 podStartE2EDuration="6.258278109s" podCreationTimestamp="2025-09-30 16:30:16 +0000 UTC" firstStartedPulling="2025-09-30 16:30:17.380117488 +0000 UTC m=+1109.393396005" lastFinishedPulling="2025-09-30 16:30:20.8003574 +0000 UTC m=+1112.813635927" observedRunningTime="2025-09-30 16:30:22.255771646 +0000 UTC m=+1114.269050183" watchObservedRunningTime="2025-09-30 16:30:22.258278109 +0000 UTC m=+1114.271556636" Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.590900 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.592324 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f340e3c8-c819-488b-808a-f610a2b60ef7" containerName="kube-state-metrics" containerID="cri-o://34a40a88436b6e90088af035949673031b8e24b425677f98cb88ecb849690d1c" gracePeriod=30 Sep 30 16:30:22 crc kubenswrapper[4796]: I0930 16:30:22.983916 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.099875 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-config-data\") pod \"919fb585-87e9-4502-90ee-a225c034f707\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.099942 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919fb585-87e9-4502-90ee-a225c034f707-logs\") pod \"919fb585-87e9-4502-90ee-a225c034f707\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.099994 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-combined-ca-bundle\") pod \"919fb585-87e9-4502-90ee-a225c034f707\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.100155 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2cfd\" (UniqueName: \"kubernetes.io/projected/919fb585-87e9-4502-90ee-a225c034f707-kube-api-access-l2cfd\") pod \"919fb585-87e9-4502-90ee-a225c034f707\" (UID: \"919fb585-87e9-4502-90ee-a225c034f707\") " Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.100649 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/919fb585-87e9-4502-90ee-a225c034f707-logs" (OuterVolumeSpecName: "logs") pod "919fb585-87e9-4502-90ee-a225c034f707" (UID: "919fb585-87e9-4502-90ee-a225c034f707"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.108227 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919fb585-87e9-4502-90ee-a225c034f707-kube-api-access-l2cfd" (OuterVolumeSpecName: "kube-api-access-l2cfd") pod "919fb585-87e9-4502-90ee-a225c034f707" (UID: "919fb585-87e9-4502-90ee-a225c034f707"). InnerVolumeSpecName "kube-api-access-l2cfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.129026 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-config-data" (OuterVolumeSpecName: "config-data") pod "919fb585-87e9-4502-90ee-a225c034f707" (UID: "919fb585-87e9-4502-90ee-a225c034f707"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.134181 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "919fb585-87e9-4502-90ee-a225c034f707" (UID: "919fb585-87e9-4502-90ee-a225c034f707"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.200849 4796 generic.go:334] "Generic (PLEG): container finished" podID="f340e3c8-c819-488b-808a-f610a2b60ef7" containerID="34a40a88436b6e90088af035949673031b8e24b425677f98cb88ecb849690d1c" exitCode=2 Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.200965 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f340e3c8-c819-488b-808a-f610a2b60ef7","Type":"ContainerDied","Data":"34a40a88436b6e90088af035949673031b8e24b425677f98cb88ecb849690d1c"} Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.201015 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f340e3c8-c819-488b-808a-f610a2b60ef7","Type":"ContainerDied","Data":"149a83074c797ec61716f77f3a3ea7b5e4c1b73dae29f799ca2c14fb83865379"} Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.201029 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="149a83074c797ec61716f77f3a3ea7b5e4c1b73dae29f799ca2c14fb83865379" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.203657 4796 generic.go:334] "Generic (PLEG): container finished" podID="919fb585-87e9-4502-90ee-a225c034f707" containerID="7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942" exitCode=0 Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.203921 4796 generic.go:334] "Generic (PLEG): container finished" podID="919fb585-87e9-4502-90ee-a225c034f707" containerID="0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e" exitCode=143 Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.203738 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.203739 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"919fb585-87e9-4502-90ee-a225c034f707","Type":"ContainerDied","Data":"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942"} Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.204042 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"919fb585-87e9-4502-90ee-a225c034f707","Type":"ContainerDied","Data":"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e"} Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.204062 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"919fb585-87e9-4502-90ee-a225c034f707","Type":"ContainerDied","Data":"c0ed9d3b940dcd989d30f8f30351ce3d763e95593cda973d9e03da47c6ef751c"} Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.204083 4796 scope.go:117] "RemoveContainer" containerID="7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.203718 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.203800 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.206250 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919fb585-87e9-4502-90ee-a225c034f707-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.206261 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919fb585-87e9-4502-90ee-a225c034f707-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.206272 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2cfd\" (UniqueName: \"kubernetes.io/projected/919fb585-87e9-4502-90ee-a225c034f707-kube-api-access-l2cfd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.253573 4796 scope.go:117] "RemoveContainer" containerID="0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.287192 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.307797 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25qtl\" (UniqueName: \"kubernetes.io/projected/f340e3c8-c819-488b-808a-f610a2b60ef7-kube-api-access-25qtl\") pod \"f340e3c8-c819-488b-808a-f610a2b60ef7\" (UID: \"f340e3c8-c819-488b-808a-f610a2b60ef7\") " Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.307804 4796 scope.go:117] "RemoveContainer" containerID="7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942" Sep 30 16:30:23 crc kubenswrapper[4796]: E0930 16:30:23.311887 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942\": container with ID starting with 7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942 not found: ID does not exist" containerID="7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.311924 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942"} err="failed to get container status \"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942\": rpc error: code = NotFound desc = could not find container \"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942\": container with ID starting with 7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942 not found: ID does not exist" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.311948 4796 scope.go:117] "RemoveContainer" containerID="0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e" Sep 30 16:30:23 crc kubenswrapper[4796]: E0930 16:30:23.312304 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e\": container with ID starting with 0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e not found: ID does not exist" containerID="0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.312357 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e"} err="failed to get container status \"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e\": rpc error: code = NotFound desc = could not find container \"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e\": container with ID starting with 0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e not found: ID does not exist" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.312394 4796 scope.go:117] "RemoveContainer" containerID="7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.312733 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942"} err="failed to get container status \"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942\": rpc error: code = NotFound desc = could not find container \"7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942\": container with ID starting with 7ed371d627b748e75ff84112636336c07c2c7f8b3051162766b7c9ae37d03942 not found: ID does not exist" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.312766 4796 scope.go:117] "RemoveContainer" containerID="0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.312863 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.313192 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f340e3c8-c819-488b-808a-f610a2b60ef7-kube-api-access-25qtl" (OuterVolumeSpecName: "kube-api-access-25qtl") pod "f340e3c8-c819-488b-808a-f610a2b60ef7" (UID: "f340e3c8-c819-488b-808a-f610a2b60ef7"). InnerVolumeSpecName "kube-api-access-25qtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.313447 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e"} err="failed to get container status \"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e\": rpc error: code = NotFound desc = could not find container \"0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e\": container with ID starting with 0b8e224d11c3c8755442d5b9194f331544845b1a47686da4f093de785a89db1e not found: ID does not exist" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320054 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:23 crc kubenswrapper[4796]: E0930 16:30:23.320482 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f340e3c8-c819-488b-808a-f610a2b60ef7" containerName="kube-state-metrics" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320499 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f340e3c8-c819-488b-808a-f610a2b60ef7" containerName="kube-state-metrics" Sep 30 16:30:23 crc kubenswrapper[4796]: E0930 16:30:23.320524 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-log" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320530 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-log" Sep 30 16:30:23 crc kubenswrapper[4796]: E0930 16:30:23.320540 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-metadata" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320546 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-metadata" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320725 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f340e3c8-c819-488b-808a-f610a2b60ef7" containerName="kube-state-metrics" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320741 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-log" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.320761 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="919fb585-87e9-4502-90ee-a225c034f707" containerName="nova-metadata-metadata" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.321728 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.326900 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.327326 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.327518 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.410801 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.410841 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-config-data\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.410991 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-logs\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.411013 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.411131 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4dn9\" (UniqueName: \"kubernetes.io/projected/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-kube-api-access-n4dn9\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.411191 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25qtl\" (UniqueName: \"kubernetes.io/projected/f340e3c8-c819-488b-808a-f610a2b60ef7-kube-api-access-25qtl\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.512699 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4dn9\" (UniqueName: \"kubernetes.io/projected/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-kube-api-access-n4dn9\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.512849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.512882 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-config-data\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.512900 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-logs\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.512926 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.513374 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-logs\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.516306 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.516470 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-config-data\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.516572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.532425 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4dn9\" (UniqueName: \"kubernetes.io/projected/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-kube-api-access-n4dn9\") pod \"nova-metadata-0\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " pod="openstack/nova-metadata-0" Sep 30 16:30:23 crc kubenswrapper[4796]: I0930 16:30:23.686744 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.157758 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:24 crc kubenswrapper[4796]: W0930 16:30:24.158895 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ee3eaa_6e30_4a84_b79b_9bb0c161d7c9.slice/crio-bce135d1733f619b233cf7ad656d04dd334f3308caf21d95a57d6b80210dd6d6 WatchSource:0}: Error finding container bce135d1733f619b233cf7ad656d04dd334f3308caf21d95a57d6b80210dd6d6: Status 404 returned error can't find the container with id bce135d1733f619b233cf7ad656d04dd334f3308caf21d95a57d6b80210dd6d6 Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.215933 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9","Type":"ContainerStarted","Data":"bce135d1733f619b233cf7ad656d04dd334f3308caf21d95a57d6b80210dd6d6"} Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.215970 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.295584 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.305919 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.320491 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.321704 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.325205 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.326895 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.329617 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.434633 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svdxf\" (UniqueName: \"kubernetes.io/projected/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-api-access-svdxf\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.434958 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.435028 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.435069 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.537215 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.537462 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svdxf\" (UniqueName: \"kubernetes.io/projected/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-api-access-svdxf\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.537492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.537535 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.541413 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.541702 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.543365 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.557378 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svdxf\" (UniqueName: \"kubernetes.io/projected/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-api-access-svdxf\") pod \"kube-state-metrics-0\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.687640 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.751323 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="919fb585-87e9-4502-90ee-a225c034f707" path="/var/lib/kubelet/pods/919fb585-87e9-4502-90ee-a225c034f707/volumes" Sep 30 16:30:24 crc kubenswrapper[4796]: I0930 16:30:24.752307 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f340e3c8-c819-488b-808a-f610a2b60ef7" path="/var/lib/kubelet/pods/f340e3c8-c819-488b-808a-f610a2b60ef7/volumes" Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.044515 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.045222 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="sg-core" containerID="cri-o://c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3" gracePeriod=30 Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.045315 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="proxy-httpd" containerID="cri-o://8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0" gracePeriod=30 Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.045315 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-notification-agent" containerID="cri-o://7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e" gracePeriod=30 Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.045188 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-central-agent" containerID="cri-o://720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56" gracePeriod=30 Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.196616 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:30:25 crc kubenswrapper[4796]: W0930 16:30:25.209609 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f48ae38_087f_474e_8e0c_6c737de6da53.slice/crio-2ee9b9bca298554a79d75bc277cbe99aac0b4f9855f9c9143d2da788bbad2348 WatchSource:0}: Error finding container 2ee9b9bca298554a79d75bc277cbe99aac0b4f9855f9c9143d2da788bbad2348: Status 404 returned error can't find the container with id 2ee9b9bca298554a79d75bc277cbe99aac0b4f9855f9c9143d2da788bbad2348 Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.226693 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9","Type":"ContainerStarted","Data":"98741f738fb61e0fd72dcef8e76b3feb462c0dd2398c2858247df66424a0f187"} Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.226739 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9","Type":"ContainerStarted","Data":"15b3402c2b32da26d3d90689aefbc3ad2dca49e46c12a533242ee45d9934a354"} Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.229490 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9f48ae38-087f-474e-8e0c-6c737de6da53","Type":"ContainerStarted","Data":"2ee9b9bca298554a79d75bc277cbe99aac0b4f9855f9c9143d2da788bbad2348"} Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.232865 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerID="c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3" exitCode=2 Sep 30 16:30:25 crc kubenswrapper[4796]: I0930 16:30:25.232904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerDied","Data":"c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3"} Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.244995 4796 generic.go:334] "Generic (PLEG): container finished" podID="f10a7487-db94-441a-856d-dff951ed56cf" containerID="a57aa13db2e426dfd044edd4989d291ba10b52a8d7e36eb6324e3600e091bc27" exitCode=0 Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.245322 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlmls" event={"ID":"f10a7487-db94-441a-856d-dff951ed56cf","Type":"ContainerDied","Data":"a57aa13db2e426dfd044edd4989d291ba10b52a8d7e36eb6324e3600e091bc27"} Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.247535 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9f48ae38-087f-474e-8e0c-6c737de6da53","Type":"ContainerStarted","Data":"52800a4bc419e5ffb650c14a5a717e854e34a5093aafc814729d1b067b5bbfe5"} Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.248168 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.251876 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerID="8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0" exitCode=0 Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.251905 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerID="720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56" exitCode=0 Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.251960 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerDied","Data":"8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0"} Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.252010 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerDied","Data":"720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56"} Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.268598 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.268580881 podStartE2EDuration="3.268580881s" podCreationTimestamp="2025-09-30 16:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:25.251096762 +0000 UTC m=+1117.264375289" watchObservedRunningTime="2025-09-30 16:30:26.268580881 +0000 UTC m=+1118.281859408" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.285421 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.827085931 podStartE2EDuration="2.285402694s" podCreationTimestamp="2025-09-30 16:30:24 +0000 UTC" firstStartedPulling="2025-09-30 16:30:25.2117314 +0000 UTC m=+1117.225009927" lastFinishedPulling="2025-09-30 16:30:25.670048153 +0000 UTC m=+1117.683326690" observedRunningTime="2025-09-30 16:30:26.281468438 +0000 UTC m=+1118.294746955" watchObservedRunningTime="2025-09-30 16:30:26.285402694 +0000 UTC m=+1118.298681221" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.557266 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.557334 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.584254 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.713492 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.713613 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.839040 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.847168 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.907228 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nnxbc"] Sep 30 16:30:26 crc kubenswrapper[4796]: I0930 16:30:26.907499 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerName="dnsmasq-dns" containerID="cri-o://9839668dab97df9c5c115feb6c77cbd558717eb50a56cb16e7ad2cf2747940fc" gracePeriod=10 Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.266466 4796 generic.go:334] "Generic (PLEG): container finished" podID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerID="9839668dab97df9c5c115feb6c77cbd558717eb50a56cb16e7ad2cf2747940fc" exitCode=0 Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.266584 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" event={"ID":"26bce871-dd26-4faf-a8a4-772a16ae7df2","Type":"ContainerDied","Data":"9839668dab97df9c5c115feb6c77cbd558717eb50a56cb16e7ad2cf2747940fc"} Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.322187 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.497368 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.597434 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czskd\" (UniqueName: \"kubernetes.io/projected/26bce871-dd26-4faf-a8a4-772a16ae7df2-kube-api-access-czskd\") pod \"26bce871-dd26-4faf-a8a4-772a16ae7df2\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.597482 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-nb\") pod \"26bce871-dd26-4faf-a8a4-772a16ae7df2\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.597580 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-sb\") pod \"26bce871-dd26-4faf-a8a4-772a16ae7df2\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.597613 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-swift-storage-0\") pod \"26bce871-dd26-4faf-a8a4-772a16ae7df2\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.597692 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-config\") pod \"26bce871-dd26-4faf-a8a4-772a16ae7df2\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.597733 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-svc\") pod \"26bce871-dd26-4faf-a8a4-772a16ae7df2\" (UID: \"26bce871-dd26-4faf-a8a4-772a16ae7df2\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.610034 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26bce871-dd26-4faf-a8a4-772a16ae7df2-kube-api-access-czskd" (OuterVolumeSpecName: "kube-api-access-czskd") pod "26bce871-dd26-4faf-a8a4-772a16ae7df2" (UID: "26bce871-dd26-4faf-a8a4-772a16ae7df2"). InnerVolumeSpecName "kube-api-access-czskd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.667654 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26bce871-dd26-4faf-a8a4-772a16ae7df2" (UID: "26bce871-dd26-4faf-a8a4-772a16ae7df2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.683442 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "26bce871-dd26-4faf-a8a4-772a16ae7df2" (UID: "26bce871-dd26-4faf-a8a4-772a16ae7df2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.694221 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26bce871-dd26-4faf-a8a4-772a16ae7df2" (UID: "26bce871-dd26-4faf-a8a4-772a16ae7df2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.696137 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26bce871-dd26-4faf-a8a4-772a16ae7df2" (UID: "26bce871-dd26-4faf-a8a4-772a16ae7df2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.707592 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czskd\" (UniqueName: \"kubernetes.io/projected/26bce871-dd26-4faf-a8a4-772a16ae7df2-kube-api-access-czskd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.707628 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.707638 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.707648 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.707661 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.723238 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-config" (OuterVolumeSpecName: "config") pod "26bce871-dd26-4faf-a8a4-772a16ae7df2" (UID: "26bce871-dd26-4faf-a8a4-772a16ae7df2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.727740 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.795199 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.795201 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.808447 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-config-data\") pod \"f10a7487-db94-441a-856d-dff951ed56cf\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.808543 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-combined-ca-bundle\") pod \"f10a7487-db94-441a-856d-dff951ed56cf\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.808615 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhwzd\" (UniqueName: \"kubernetes.io/projected/f10a7487-db94-441a-856d-dff951ed56cf-kube-api-access-zhwzd\") pod \"f10a7487-db94-441a-856d-dff951ed56cf\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.808791 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-scripts\") pod \"f10a7487-db94-441a-856d-dff951ed56cf\" (UID: \"f10a7487-db94-441a-856d-dff951ed56cf\") " Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.809262 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bce871-dd26-4faf-a8a4-772a16ae7df2-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.814162 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10a7487-db94-441a-856d-dff951ed56cf-kube-api-access-zhwzd" (OuterVolumeSpecName: "kube-api-access-zhwzd") pod "f10a7487-db94-441a-856d-dff951ed56cf" (UID: "f10a7487-db94-441a-856d-dff951ed56cf"). InnerVolumeSpecName "kube-api-access-zhwzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.814835 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-scripts" (OuterVolumeSpecName: "scripts") pod "f10a7487-db94-441a-856d-dff951ed56cf" (UID: "f10a7487-db94-441a-856d-dff951ed56cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.855670 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-config-data" (OuterVolumeSpecName: "config-data") pod "f10a7487-db94-441a-856d-dff951ed56cf" (UID: "f10a7487-db94-441a-856d-dff951ed56cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.868600 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f10a7487-db94-441a-856d-dff951ed56cf" (UID: "f10a7487-db94-441a-856d-dff951ed56cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.912064 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhwzd\" (UniqueName: \"kubernetes.io/projected/f10a7487-db94-441a-856d-dff951ed56cf-kube-api-access-zhwzd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.912364 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.912480 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:27 crc kubenswrapper[4796]: I0930 16:30:27.912578 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10a7487-db94-441a-856d-dff951ed56cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.277734 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" event={"ID":"26bce871-dd26-4faf-a8a4-772a16ae7df2","Type":"ContainerDied","Data":"6ec5467dd9031a6cdc415c942a44bda1b566f8f78edcbc67f20e9138d7b2df4d"} Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.278040 4796 scope.go:117] "RemoveContainer" containerID="9839668dab97df9c5c115feb6c77cbd558717eb50a56cb16e7ad2cf2747940fc" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.278200 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nnxbc" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.282205 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlmls" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.282175 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlmls" event={"ID":"f10a7487-db94-441a-856d-dff951ed56cf","Type":"ContainerDied","Data":"7c13497c2b2bdb3ae6766feb417de9af623f0b310fc0889a9cdab32d5466ed70"} Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.282249 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c13497c2b2bdb3ae6766feb417de9af623f0b310fc0889a9cdab32d5466ed70" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.305590 4796 scope.go:117] "RemoveContainer" containerID="981589e343b339b655a98d3e8e67ab8e93106995711ce81b59076ab7126c3454" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.339671 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nnxbc"] Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.350123 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nnxbc"] Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.427952 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.428237 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-log" containerID="cri-o://4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819" gracePeriod=30 Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.428292 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-api" containerID="cri-o://c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50" gracePeriod=30 Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.435302 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.448777 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.448992 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-log" containerID="cri-o://15b3402c2b32da26d3d90689aefbc3ad2dca49e46c12a533242ee45d9934a354" gracePeriod=30 Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.450010 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-metadata" containerID="cri-o://98741f738fb61e0fd72dcef8e76b3feb462c0dd2398c2858247df66424a0f187" gracePeriod=30 Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.688050 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.688090 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.747192 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" path="/var/lib/kubelet/pods/26bce871-dd26-4faf-a8a4-772a16ae7df2/volumes" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.778267 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.933882 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-run-httpd\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.933949 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29h8q\" (UniqueName: \"kubernetes.io/projected/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-kube-api-access-29h8q\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.934005 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-combined-ca-bundle\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.934114 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-config-data\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.934162 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-scripts\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.934189 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-log-httpd\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.934244 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-sg-core-conf-yaml\") pod \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\" (UID: \"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2\") " Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.935245 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.937617 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.944141 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-kube-api-access-29h8q" (OuterVolumeSpecName: "kube-api-access-29h8q") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "kube-api-access-29h8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.944163 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-scripts" (OuterVolumeSpecName: "scripts") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:28 crc kubenswrapper[4796]: I0930 16:30:28.996192 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.037998 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.038049 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29h8q\" (UniqueName: \"kubernetes.io/projected/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-kube-api-access-29h8q\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.038062 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.038075 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.038086 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.048322 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.067942 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-config-data" (OuterVolumeSpecName: "config-data") pod "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" (UID: "8a6c845a-7bcd-4035-9255-f8c3bb92a1e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.140305 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.140344 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.291666 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerID="7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e" exitCode=0 Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.291954 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerDied","Data":"7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e"} Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.291996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a6c845a-7bcd-4035-9255-f8c3bb92a1e2","Type":"ContainerDied","Data":"e0e5b0c8f4d0100ec1807f2d6fac250b17ea0282f8e9fa777c55b60247e7a519"} Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.292013 4796 scope.go:117] "RemoveContainer" containerID="8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.292111 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.309294 4796 generic.go:334] "Generic (PLEG): container finished" podID="5357c107-43c4-46a5-b422-e1caca9bf032" containerID="4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819" exitCode=143 Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.309385 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5357c107-43c4-46a5-b422-e1caca9bf032","Type":"ContainerDied","Data":"4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819"} Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.321574 4796 generic.go:334] "Generic (PLEG): container finished" podID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerID="98741f738fb61e0fd72dcef8e76b3feb462c0dd2398c2858247df66424a0f187" exitCode=0 Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.321601 4796 generic.go:334] "Generic (PLEG): container finished" podID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerID="15b3402c2b32da26d3d90689aefbc3ad2dca49e46c12a533242ee45d9934a354" exitCode=143 Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.321764 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" containerName="nova-scheduler-scheduler" containerID="cri-o://a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" gracePeriod=30 Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.321885 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9","Type":"ContainerDied","Data":"98741f738fb61e0fd72dcef8e76b3feb462c0dd2398c2858247df66424a0f187"} Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.321913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9","Type":"ContainerDied","Data":"15b3402c2b32da26d3d90689aefbc3ad2dca49e46c12a533242ee45d9934a354"} Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.357495 4796 scope.go:117] "RemoveContainer" containerID="c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.372110 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.387568 4796 scope.go:117] "RemoveContainer" containerID="7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.393482 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.422043 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423005 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-notification-agent" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423079 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-notification-agent" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423176 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerName="dnsmasq-dns" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423230 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerName="dnsmasq-dns" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423319 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="sg-core" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423373 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="sg-core" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423431 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-central-agent" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423483 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-central-agent" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423565 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerName="init" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423663 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerName="init" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423737 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10a7487-db94-441a-856d-dff951ed56cf" containerName="nova-manage" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423794 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10a7487-db94-441a-856d-dff951ed56cf" containerName="nova-manage" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.423875 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="proxy-httpd" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.423928 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="proxy-httpd" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.424384 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-notification-agent" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.424783 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="26bce871-dd26-4faf-a8a4-772a16ae7df2" containerName="dnsmasq-dns" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.424886 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="ceilometer-central-agent" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.425065 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10a7487-db94-441a-856d-dff951ed56cf" containerName="nova-manage" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.425114 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="sg-core" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.425134 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" containerName="proxy-httpd" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.433338 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.437352 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.437563 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.437773 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.453541 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.476703 4796 scope.go:117] "RemoveContainer" containerID="720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.518550 4796 scope.go:117] "RemoveContainer" containerID="8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.520434 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0\": container with ID starting with 8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0 not found: ID does not exist" containerID="8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.520582 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0"} err="failed to get container status \"8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0\": rpc error: code = NotFound desc = could not find container \"8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0\": container with ID starting with 8661ee9ce785795f89faeba9e24004ee88230d3fa3ec439ac43246e48fc59ec0 not found: ID does not exist" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.520722 4796 scope.go:117] "RemoveContainer" containerID="c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.521042 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3\": container with ID starting with c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3 not found: ID does not exist" containerID="c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.521082 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3"} err="failed to get container status \"c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3\": rpc error: code = NotFound desc = could not find container \"c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3\": container with ID starting with c17411ab32a2aad57d3ba8fbbffb973e4940b5e0f75c7e35665d493dc92273e3 not found: ID does not exist" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.521109 4796 scope.go:117] "RemoveContainer" containerID="7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.521515 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e\": container with ID starting with 7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e not found: ID does not exist" containerID="7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.521534 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e"} err="failed to get container status \"7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e\": rpc error: code = NotFound desc = could not find container \"7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e\": container with ID starting with 7f93e92f905b90afa026a6f4e612db4345937a939924fb7e24d20b56fb7d8a1e not found: ID does not exist" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.521547 4796 scope.go:117] "RemoveContainer" containerID="720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56" Sep 30 16:30:29 crc kubenswrapper[4796]: E0930 16:30:29.521731 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56\": container with ID starting with 720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56 not found: ID does not exist" containerID="720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.521752 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56"} err="failed to get container status \"720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56\": rpc error: code = NotFound desc = could not find container \"720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56\": container with ID starting with 720fe7b442638c183c5a929d7f1e0b94bde070999c71b7d227b222387a071d56 not found: ID does not exist" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.558952 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-scripts\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559199 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559300 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-run-httpd\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559393 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-config-data\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559567 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zszsv\" (UniqueName: \"kubernetes.io/projected/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-kube-api-access-zszsv\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559687 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559848 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-log-httpd\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.559997 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.608549 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662334 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-config-data\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662455 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zszsv\" (UniqueName: \"kubernetes.io/projected/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-kube-api-access-zszsv\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662501 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662549 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-log-httpd\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662581 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-scripts\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662685 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.662711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-run-httpd\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.663131 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-run-httpd\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.664055 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-log-httpd\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.668893 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-config-data\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.669357 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.669817 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-scripts\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.670737 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.682847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.691416 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zszsv\" (UniqueName: \"kubernetes.io/projected/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-kube-api-access-zszsv\") pod \"ceilometer-0\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.763885 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-logs\") pod \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.763930 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-nova-metadata-tls-certs\") pod \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.763949 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4dn9\" (UniqueName: \"kubernetes.io/projected/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-kube-api-access-n4dn9\") pod \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.764046 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-config-data\") pod \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.764099 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-combined-ca-bundle\") pod \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\" (UID: \"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9\") " Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.764261 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-logs" (OuterVolumeSpecName: "logs") pod "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" (UID: "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.765106 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.767526 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-kube-api-access-n4dn9" (OuterVolumeSpecName: "kube-api-access-n4dn9") pod "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" (UID: "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9"). InnerVolumeSpecName "kube-api-access-n4dn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.795449 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.799087 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" (UID: "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.809079 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-config-data" (OuterVolumeSpecName: "config-data") pod "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" (UID: "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.820634 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" (UID: "93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.866538 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.866574 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.866585 4796 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:29 crc kubenswrapper[4796]: I0930 16:30:29.866594 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4dn9\" (UniqueName: \"kubernetes.io/projected/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9-kube-api-access-n4dn9\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.280112 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.336829 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9","Type":"ContainerDied","Data":"bce135d1733f619b233cf7ad656d04dd334f3308caf21d95a57d6b80210dd6d6"} Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.336904 4796 scope.go:117] "RemoveContainer" containerID="98741f738fb61e0fd72dcef8e76b3feb462c0dd2398c2858247df66424a0f187" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.337077 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.340913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerStarted","Data":"e1b40130d6a6a9c14cef0d47643b2c356c8f05fa76d3adad9c7fcbd289163ca9"} Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.343238 4796 generic.go:334] "Generic (PLEG): container finished" podID="800b5cc2-8806-4298-802a-7bef00b36a4c" containerID="d1e335a5c446dfe36da0709c7daadfd01ef39f4982921030c7fc75692148aa8e" exitCode=0 Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.343419 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jpscz" event={"ID":"800b5cc2-8806-4298-802a-7bef00b36a4c","Type":"ContainerDied","Data":"d1e335a5c446dfe36da0709c7daadfd01ef39f4982921030c7fc75692148aa8e"} Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.364915 4796 scope.go:117] "RemoveContainer" containerID="15b3402c2b32da26d3d90689aefbc3ad2dca49e46c12a533242ee45d9934a354" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.412026 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.427178 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.437129 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:30 crc kubenswrapper[4796]: E0930 16:30:30.437572 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-log" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.437590 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-log" Sep 30 16:30:30 crc kubenswrapper[4796]: E0930 16:30:30.437616 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-metadata" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.437624 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-metadata" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.437848 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-log" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.437887 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" containerName="nova-metadata-metadata" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.439111 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.441539 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.450457 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.450938 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.477843 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whb5h\" (UniqueName: \"kubernetes.io/projected/7e16f45a-71c1-411e-9084-c950519e768d-kube-api-access-whb5h\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.477918 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-config-data\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.477967 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.478033 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e16f45a-71c1-411e-9084-c950519e768d-logs\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.478114 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.579033 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.579141 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whb5h\" (UniqueName: \"kubernetes.io/projected/7e16f45a-71c1-411e-9084-c950519e768d-kube-api-access-whb5h\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.579187 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-config-data\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.579763 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.579941 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e16f45a-71c1-411e-9084-c950519e768d-logs\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.580505 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e16f45a-71c1-411e-9084-c950519e768d-logs\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.586854 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.589600 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.590369 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-config-data\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.606441 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whb5h\" (UniqueName: \"kubernetes.io/projected/7e16f45a-71c1-411e-9084-c950519e768d-kube-api-access-whb5h\") pod \"nova-metadata-0\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " pod="openstack/nova-metadata-0" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.745176 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6c845a-7bcd-4035-9255-f8c3bb92a1e2" path="/var/lib/kubelet/pods/8a6c845a-7bcd-4035-9255-f8c3bb92a1e2/volumes" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.745905 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9" path="/var/lib/kubelet/pods/93ee3eaa-6e30-4a84-b79b-9bb0c161d7c9/volumes" Sep 30 16:30:30 crc kubenswrapper[4796]: I0930 16:30:30.759769 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.096838 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.097252 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.097304 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.098224 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f24d81f5faa8f6182a7d380f63e2c090bd17e98f5259e863fa184efdaf83296f"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.098291 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://f24d81f5faa8f6182a7d380f63e2c090bd17e98f5259e863fa184efdaf83296f" gracePeriod=600 Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.221478 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.373042 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerStarted","Data":"c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632"} Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.376508 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e16f45a-71c1-411e-9084-c950519e768d","Type":"ContainerStarted","Data":"32d4d3ca89b37a8cabd21e2fe36a8c30348d629e3ca7a0bd0073ecd6685a5c7d"} Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.379392 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="f24d81f5faa8f6182a7d380f63e2c090bd17e98f5259e863fa184efdaf83296f" exitCode=0 Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.379446 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"f24d81f5faa8f6182a7d380f63e2c090bd17e98f5259e863fa184efdaf83296f"} Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.379473 4796 scope.go:117] "RemoveContainer" containerID="f80108aae73ea16e4bbeae96cdafd6d40fe314e16173a204b191d54c90a06ba4" Sep 30 16:30:31 crc kubenswrapper[4796]: E0930 16:30:31.557641 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 16:30:31 crc kubenswrapper[4796]: E0930 16:30:31.559126 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 16:30:31 crc kubenswrapper[4796]: E0930 16:30:31.560949 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 16:30:31 crc kubenswrapper[4796]: E0930 16:30:31.561000 4796 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" containerName="nova-scheduler-scheduler" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.636392 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.810004 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-config-data\") pod \"800b5cc2-8806-4298-802a-7bef00b36a4c\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.810347 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56rkt\" (UniqueName: \"kubernetes.io/projected/800b5cc2-8806-4298-802a-7bef00b36a4c-kube-api-access-56rkt\") pod \"800b5cc2-8806-4298-802a-7bef00b36a4c\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.810404 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-scripts\") pod \"800b5cc2-8806-4298-802a-7bef00b36a4c\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.810460 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-combined-ca-bundle\") pod \"800b5cc2-8806-4298-802a-7bef00b36a4c\" (UID: \"800b5cc2-8806-4298-802a-7bef00b36a4c\") " Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.819102 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/800b5cc2-8806-4298-802a-7bef00b36a4c-kube-api-access-56rkt" (OuterVolumeSpecName: "kube-api-access-56rkt") pod "800b5cc2-8806-4298-802a-7bef00b36a4c" (UID: "800b5cc2-8806-4298-802a-7bef00b36a4c"). InnerVolumeSpecName "kube-api-access-56rkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.821632 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-scripts" (OuterVolumeSpecName: "scripts") pod "800b5cc2-8806-4298-802a-7bef00b36a4c" (UID: "800b5cc2-8806-4298-802a-7bef00b36a4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.851800 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-config-data" (OuterVolumeSpecName: "config-data") pod "800b5cc2-8806-4298-802a-7bef00b36a4c" (UID: "800b5cc2-8806-4298-802a-7bef00b36a4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.852215 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "800b5cc2-8806-4298-802a-7bef00b36a4c" (UID: "800b5cc2-8806-4298-802a-7bef00b36a4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.912327 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.912364 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.912373 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800b5cc2-8806-4298-802a-7bef00b36a4c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:31 crc kubenswrapper[4796]: I0930 16:30:31.912383 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56rkt\" (UniqueName: \"kubernetes.io/projected/800b5cc2-8806-4298-802a-7bef00b36a4c-kube-api-access-56rkt\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.419398 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"7b7ed0c28fbd614f1b48f9a4c2dd1b94dd52891d55126b94c8724c34b2b684f2"} Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.431595 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerStarted","Data":"c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45"} Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.433478 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jpscz" event={"ID":"800b5cc2-8806-4298-802a-7bef00b36a4c","Type":"ContainerDied","Data":"e8eb7a4f52c537b8bc32cbbb975b7149117a7634f3a59708eca8ee9fa8b08b01"} Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.433507 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8eb7a4f52c537b8bc32cbbb975b7149117a7634f3a59708eca8ee9fa8b08b01" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.433569 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jpscz" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.441165 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 16:30:32 crc kubenswrapper[4796]: E0930 16:30:32.448875 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800b5cc2-8806-4298-802a-7bef00b36a4c" containerName="nova-cell1-conductor-db-sync" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.448909 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="800b5cc2-8806-4298-802a-7bef00b36a4c" containerName="nova-cell1-conductor-db-sync" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.449613 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="800b5cc2-8806-4298-802a-7bef00b36a4c" containerName="nova-cell1-conductor-db-sync" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.450649 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.453551 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.458528 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e16f45a-71c1-411e-9084-c950519e768d","Type":"ContainerStarted","Data":"9e0c4fa9f51e60247a9f5b85ab6550c10ea58c4fae85253c842c7f6f8634a34f"} Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.458589 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e16f45a-71c1-411e-9084-c950519e768d","Type":"ContainerStarted","Data":"8923fde2eba2d93d4cf942b68a5fadeec1b42fcb138fe20b5d3238a83dffc15f"} Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.476522 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.521455 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.521434368 podStartE2EDuration="2.521434368s" podCreationTimestamp="2025-09-30 16:30:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:32.505731448 +0000 UTC m=+1124.519009975" watchObservedRunningTime="2025-09-30 16:30:32.521434368 +0000 UTC m=+1124.534712895" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.532633 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5b132cf-1062-45b8-a1b2-8626de3f640d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.532709 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52gcn\" (UniqueName: \"kubernetes.io/projected/d5b132cf-1062-45b8-a1b2-8626de3f640d-kube-api-access-52gcn\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.532943 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5b132cf-1062-45b8-a1b2-8626de3f640d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.633838 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5b132cf-1062-45b8-a1b2-8626de3f640d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.633892 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52gcn\" (UniqueName: \"kubernetes.io/projected/d5b132cf-1062-45b8-a1b2-8626de3f640d-kube-api-access-52gcn\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.633936 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5b132cf-1062-45b8-a1b2-8626de3f640d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.639065 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5b132cf-1062-45b8-a1b2-8626de3f640d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.640847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5b132cf-1062-45b8-a1b2-8626de3f640d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.661610 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52gcn\" (UniqueName: \"kubernetes.io/projected/d5b132cf-1062-45b8-a1b2-8626de3f640d-kube-api-access-52gcn\") pod \"nova-cell1-conductor-0\" (UID: \"d5b132cf-1062-45b8-a1b2-8626de3f640d\") " pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:32 crc kubenswrapper[4796]: I0930 16:30:32.786022 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.280335 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.435094 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.471614 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerStarted","Data":"64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777"} Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.473347 4796 generic.go:334] "Generic (PLEG): container finished" podID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" exitCode=0 Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.473389 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3","Type":"ContainerDied","Data":"a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932"} Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.473412 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.473432 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3","Type":"ContainerDied","Data":"579249f88007949d974e24497896e264b1ed3054caf2118befebe1cadd1745cf"} Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.473451 4796 scope.go:117] "RemoveContainer" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.476426 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d5b132cf-1062-45b8-a1b2-8626de3f640d","Type":"ContainerStarted","Data":"62ada8f169bea9bb67f026a201205a63a76bbec6e8b8deba531407cb6ba49374"} Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.491047 4796 scope.go:117] "RemoveContainer" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" Sep 30 16:30:33 crc kubenswrapper[4796]: E0930 16:30:33.492368 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932\": container with ID starting with a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932 not found: ID does not exist" containerID="a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.492403 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932"} err="failed to get container status \"a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932\": rpc error: code = NotFound desc = could not find container \"a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932\": container with ID starting with a724f56522cf5fc3901935220d80d25a696cb7be9d1ad8dd5927723856d86932 not found: ID does not exist" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.547399 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-config-data\") pod \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.547530 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-combined-ca-bundle\") pod \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.547606 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljjsc\" (UniqueName: \"kubernetes.io/projected/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-kube-api-access-ljjsc\") pod \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\" (UID: \"e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3\") " Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.553084 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-kube-api-access-ljjsc" (OuterVolumeSpecName: "kube-api-access-ljjsc") pod "e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" (UID: "e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3"). InnerVolumeSpecName "kube-api-access-ljjsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.576757 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" (UID: "e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.579123 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-config-data" (OuterVolumeSpecName: "config-data") pod "e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" (UID: "e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.649176 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.649224 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.649237 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljjsc\" (UniqueName: \"kubernetes.io/projected/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3-kube-api-access-ljjsc\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.818736 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.827526 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.837498 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:33 crc kubenswrapper[4796]: E0930 16:30:33.837956 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" containerName="nova-scheduler-scheduler" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.837998 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" containerName="nova-scheduler-scheduler" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.838196 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" containerName="nova-scheduler-scheduler" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.838829 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.840753 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.845720 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.957148 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5jhf\" (UniqueName: \"kubernetes.io/projected/0d226c88-71f5-4831-8576-4c8d02c0fb30-kube-api-access-f5jhf\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.957332 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:33 crc kubenswrapper[4796]: I0930 16:30:33.957399 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-config-data\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.058672 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5jhf\" (UniqueName: \"kubernetes.io/projected/0d226c88-71f5-4831-8576-4c8d02c0fb30-kube-api-access-f5jhf\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.058779 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.058839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-config-data\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.065090 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.065634 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-config-data\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.079532 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5jhf\" (UniqueName: \"kubernetes.io/projected/0d226c88-71f5-4831-8576-4c8d02c0fb30-kube-api-access-f5jhf\") pod \"nova-scheduler-0\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.179316 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.375539 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.497417 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d5b132cf-1062-45b8-a1b2-8626de3f640d","Type":"ContainerStarted","Data":"4730025b7475a2e3c8aa5186e8a6233c3daff007c48915513dfe94c87064ff7b"} Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.499168 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.513853 4796 generic.go:334] "Generic (PLEG): container finished" podID="5357c107-43c4-46a5-b422-e1caca9bf032" containerID="c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50" exitCode=0 Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.513898 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.513925 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5357c107-43c4-46a5-b422-e1caca9bf032","Type":"ContainerDied","Data":"c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50"} Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.513958 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5357c107-43c4-46a5-b422-e1caca9bf032","Type":"ContainerDied","Data":"e42fec612e50a69c843ba7696eb57a8cda5d126bcd23edb8c50fd67534f90b44"} Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.513993 4796 scope.go:117] "RemoveContainer" containerID="c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.527687 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.527666116 podStartE2EDuration="2.527666116s" podCreationTimestamp="2025-09-30 16:30:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:34.523260397 +0000 UTC m=+1126.536538924" watchObservedRunningTime="2025-09-30 16:30:34.527666116 +0000 UTC m=+1126.540944643" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.575578 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w9c6\" (UniqueName: \"kubernetes.io/projected/5357c107-43c4-46a5-b422-e1caca9bf032-kube-api-access-4w9c6\") pod \"5357c107-43c4-46a5-b422-e1caca9bf032\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.575740 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-config-data\") pod \"5357c107-43c4-46a5-b422-e1caca9bf032\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.575773 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-combined-ca-bundle\") pod \"5357c107-43c4-46a5-b422-e1caca9bf032\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.575836 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5357c107-43c4-46a5-b422-e1caca9bf032-logs\") pod \"5357c107-43c4-46a5-b422-e1caca9bf032\" (UID: \"5357c107-43c4-46a5-b422-e1caca9bf032\") " Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.577711 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5357c107-43c4-46a5-b422-e1caca9bf032-logs" (OuterVolumeSpecName: "logs") pod "5357c107-43c4-46a5-b422-e1caca9bf032" (UID: "5357c107-43c4-46a5-b422-e1caca9bf032"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.584160 4796 scope.go:117] "RemoveContainer" containerID="4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.586210 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5357c107-43c4-46a5-b422-e1caca9bf032-kube-api-access-4w9c6" (OuterVolumeSpecName: "kube-api-access-4w9c6") pod "5357c107-43c4-46a5-b422-e1caca9bf032" (UID: "5357c107-43c4-46a5-b422-e1caca9bf032"). InnerVolumeSpecName "kube-api-access-4w9c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.627107 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5357c107-43c4-46a5-b422-e1caca9bf032" (UID: "5357c107-43c4-46a5-b422-e1caca9bf032"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.629582 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-config-data" (OuterVolumeSpecName: "config-data") pod "5357c107-43c4-46a5-b422-e1caca9bf032" (UID: "5357c107-43c4-46a5-b422-e1caca9bf032"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.645519 4796 scope.go:117] "RemoveContainer" containerID="c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50" Sep 30 16:30:34 crc kubenswrapper[4796]: E0930 16:30:34.652414 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50\": container with ID starting with c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50 not found: ID does not exist" containerID="c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.652477 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50"} err="failed to get container status \"c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50\": rpc error: code = NotFound desc = could not find container \"c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50\": container with ID starting with c62da8b3b6f3a788f65936b1a44b78622181046d6228d90bbfc84071b99cca50 not found: ID does not exist" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.652541 4796 scope.go:117] "RemoveContainer" containerID="4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819" Sep 30 16:30:34 crc kubenswrapper[4796]: E0930 16:30:34.652912 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819\": container with ID starting with 4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819 not found: ID does not exist" containerID="4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.652934 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819"} err="failed to get container status \"4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819\": rpc error: code = NotFound desc = could not find container \"4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819\": container with ID starting with 4e1c3726e905d1938ffb1a8255f02110abe1649274ca793035ed0fba9bb5e819 not found: ID does not exist" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.658128 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.678092 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w9c6\" (UniqueName: \"kubernetes.io/projected/5357c107-43c4-46a5-b422-e1caca9bf032-kube-api-access-4w9c6\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.678121 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.678132 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5357c107-43c4-46a5-b422-e1caca9bf032-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.678143 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5357c107-43c4-46a5-b422-e1caca9bf032-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.707386 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.758022 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3" path="/var/lib/kubelet/pods/e77aa5c1-9906-46b3-8b0c-81fbf4a23fc3/volumes" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.844078 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.861887 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.884691 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:34 crc kubenswrapper[4796]: E0930 16:30:34.894443 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-log" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.894476 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-log" Sep 30 16:30:34 crc kubenswrapper[4796]: E0930 16:30:34.894514 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-api" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.894521 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-api" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.894945 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-log" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.894993 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" containerName="nova-api-api" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.900062 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.902825 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.904109 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.993058 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pws5\" (UniqueName: \"kubernetes.io/projected/4ee982da-2a4e-4e01-8315-acf0dd590839-kube-api-access-7pws5\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.993147 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.993272 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee982da-2a4e-4e01-8315-acf0dd590839-logs\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:34 crc kubenswrapper[4796]: I0930 16:30:34.993330 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-config-data\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.094572 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-config-data\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.094667 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pws5\" (UniqueName: \"kubernetes.io/projected/4ee982da-2a4e-4e01-8315-acf0dd590839-kube-api-access-7pws5\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.094693 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.094779 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee982da-2a4e-4e01-8315-acf0dd590839-logs\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.095160 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee982da-2a4e-4e01-8315-acf0dd590839-logs\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.099622 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-config-data\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.101435 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.114393 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pws5\" (UniqueName: \"kubernetes.io/projected/4ee982da-2a4e-4e01-8315-acf0dd590839-kube-api-access-7pws5\") pod \"nova-api-0\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.239599 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.539045 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d226c88-71f5-4831-8576-4c8d02c0fb30","Type":"ContainerStarted","Data":"f5d2055e0daa8a6c32500446e823ae401143de578c663dc0332c12e79c19b0ae"} Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.539635 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d226c88-71f5-4831-8576-4c8d02c0fb30","Type":"ContainerStarted","Data":"db691c3867efd2379c37fc22fb1c8f3deda15a55119cadd9193fd22e12f88984"} Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.563464 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.56344783 podStartE2EDuration="2.56344783s" podCreationTimestamp="2025-09-30 16:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:35.559834924 +0000 UTC m=+1127.573113461" watchObservedRunningTime="2025-09-30 16:30:35.56344783 +0000 UTC m=+1127.576726357" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.749780 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.760658 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 16:30:35 crc kubenswrapper[4796]: I0930 16:30:35.761608 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.553769 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ee982da-2a4e-4e01-8315-acf0dd590839","Type":"ContainerStarted","Data":"9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2"} Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.554039 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ee982da-2a4e-4e01-8315-acf0dd590839","Type":"ContainerStarted","Data":"e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f"} Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.554050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ee982da-2a4e-4e01-8315-acf0dd590839","Type":"ContainerStarted","Data":"5fe808b61b62eee5bb37314ab7690635124f5ec060a2235d50ebd44adf9343c8"} Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.557871 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerStarted","Data":"6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3"} Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.584855 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.584826263 podStartE2EDuration="2.584826263s" podCreationTimestamp="2025-09-30 16:30:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:36.577251191 +0000 UTC m=+1128.590529718" watchObservedRunningTime="2025-09-30 16:30:36.584826263 +0000 UTC m=+1128.598104810" Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.613811 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.938698635 podStartE2EDuration="7.613793661s" podCreationTimestamp="2025-09-30 16:30:29 +0000 UTC" firstStartedPulling="2025-09-30 16:30:30.291187114 +0000 UTC m=+1122.304465641" lastFinishedPulling="2025-09-30 16:30:35.96628214 +0000 UTC m=+1127.979560667" observedRunningTime="2025-09-30 16:30:36.604092657 +0000 UTC m=+1128.617371194" watchObservedRunningTime="2025-09-30 16:30:36.613793661 +0000 UTC m=+1128.627072178" Sep 30 16:30:36 crc kubenswrapper[4796]: I0930 16:30:36.754727 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5357c107-43c4-46a5-b422-e1caca9bf032" path="/var/lib/kubelet/pods/5357c107-43c4-46a5-b422-e1caca9bf032/volumes" Sep 30 16:30:37 crc kubenswrapper[4796]: I0930 16:30:37.567546 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:30:39 crc kubenswrapper[4796]: I0930 16:30:39.180606 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 16:30:40 crc kubenswrapper[4796]: I0930 16:30:40.760274 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 16:30:40 crc kubenswrapper[4796]: I0930 16:30:40.760319 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 16:30:41 crc kubenswrapper[4796]: I0930 16:30:41.777248 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 16:30:41 crc kubenswrapper[4796]: I0930 16:30:41.777317 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 16:30:42 crc kubenswrapper[4796]: I0930 16:30:42.816237 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 30 16:30:44 crc kubenswrapper[4796]: I0930 16:30:44.180423 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 16:30:44 crc kubenswrapper[4796]: I0930 16:30:44.214671 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 16:30:44 crc kubenswrapper[4796]: I0930 16:30:44.698374 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 16:30:45 crc kubenswrapper[4796]: I0930 16:30:45.240431 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 16:30:45 crc kubenswrapper[4796]: I0930 16:30:45.240786 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 16:30:46 crc kubenswrapper[4796]: I0930 16:30:46.322197 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 16:30:46 crc kubenswrapper[4796]: I0930 16:30:46.322243 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 16:30:50 crc kubenswrapper[4796]: I0930 16:30:50.774966 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 16:30:50 crc kubenswrapper[4796]: I0930 16:30:50.780244 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 16:30:50 crc kubenswrapper[4796]: I0930 16:30:50.784857 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 16:30:51 crc kubenswrapper[4796]: I0930 16:30:51.738987 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.601576 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.659104 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mmw2\" (UniqueName: \"kubernetes.io/projected/720f2613-b6eb-43df-95a2-d05f980bb0c0-kube-api-access-5mmw2\") pod \"720f2613-b6eb-43df-95a2-d05f980bb0c0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.659217 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-config-data\") pod \"720f2613-b6eb-43df-95a2-d05f980bb0c0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.659328 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-combined-ca-bundle\") pod \"720f2613-b6eb-43df-95a2-d05f980bb0c0\" (UID: \"720f2613-b6eb-43df-95a2-d05f980bb0c0\") " Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.667300 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/720f2613-b6eb-43df-95a2-d05f980bb0c0-kube-api-access-5mmw2" (OuterVolumeSpecName: "kube-api-access-5mmw2") pod "720f2613-b6eb-43df-95a2-d05f980bb0c0" (UID: "720f2613-b6eb-43df-95a2-d05f980bb0c0"). InnerVolumeSpecName "kube-api-access-5mmw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.689439 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "720f2613-b6eb-43df-95a2-d05f980bb0c0" (UID: "720f2613-b6eb-43df-95a2-d05f980bb0c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.700606 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-config-data" (OuterVolumeSpecName: "config-data") pod "720f2613-b6eb-43df-95a2-d05f980bb0c0" (UID: "720f2613-b6eb-43df-95a2-d05f980bb0c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.755464 4796 generic.go:334] "Generic (PLEG): container finished" podID="720f2613-b6eb-43df-95a2-d05f980bb0c0" containerID="33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7" exitCode=137 Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.755872 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.755877 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"720f2613-b6eb-43df-95a2-d05f980bb0c0","Type":"ContainerDied","Data":"33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7"} Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.756044 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"720f2613-b6eb-43df-95a2-d05f980bb0c0","Type":"ContainerDied","Data":"9ae8c12c6553751e748202aad4d487b2137c98a19705db11ae3115ac15f872e1"} Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.756138 4796 scope.go:117] "RemoveContainer" containerID="33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.761623 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mmw2\" (UniqueName: \"kubernetes.io/projected/720f2613-b6eb-43df-95a2-d05f980bb0c0-kube-api-access-5mmw2\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.761656 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.761670 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720f2613-b6eb-43df-95a2-d05f980bb0c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.793086 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.797579 4796 scope.go:117] "RemoveContainer" containerID="33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7" Sep 30 16:30:52 crc kubenswrapper[4796]: E0930 16:30:52.798068 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7\": container with ID starting with 33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7 not found: ID does not exist" containerID="33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.798126 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7"} err="failed to get container status \"33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7\": rpc error: code = NotFound desc = could not find container \"33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7\": container with ID starting with 33d7429b40a2e2d9785866434d656ebb4b0a66608db0f5c0f6217bba9752cde7 not found: ID does not exist" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.815140 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.831595 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:52 crc kubenswrapper[4796]: E0930 16:30:52.832154 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720f2613-b6eb-43df-95a2-d05f980bb0c0" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.832181 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="720f2613-b6eb-43df-95a2-d05f980bb0c0" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.832449 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="720f2613-b6eb-43df-95a2-d05f980bb0c0" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.835345 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.841321 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.842382 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.842915 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.873160 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.968947 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.969024 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnwl9\" (UniqueName: \"kubernetes.io/projected/8a76e9ab-55ec-4394-94b3-668b7b146c81-kube-api-access-gnwl9\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.969048 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.969090 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:52 crc kubenswrapper[4796]: I0930 16:30:52.969152 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.071197 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.071275 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnwl9\" (UniqueName: \"kubernetes.io/projected/8a76e9ab-55ec-4394-94b3-668b7b146c81-kube-api-access-gnwl9\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.071306 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.071386 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.071515 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.075320 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.075328 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.076077 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.076866 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a76e9ab-55ec-4394-94b3-668b7b146c81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.091106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnwl9\" (UniqueName: \"kubernetes.io/projected/8a76e9ab-55ec-4394-94b3-668b7b146c81-kube-api-access-gnwl9\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a76e9ab-55ec-4394-94b3-668b7b146c81\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.167850 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:53 crc kubenswrapper[4796]: W0930 16:30:53.740736 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a76e9ab_55ec_4394_94b3_668b7b146c81.slice/crio-b04152707f925534b2a04f0c82b5b6ac4d9febf97cb12cf53faef63e23ed4e81 WatchSource:0}: Error finding container b04152707f925534b2a04f0c82b5b6ac4d9febf97cb12cf53faef63e23ed4e81: Status 404 returned error can't find the container with id b04152707f925534b2a04f0c82b5b6ac4d9febf97cb12cf53faef63e23ed4e81 Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.742130 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 16:30:53 crc kubenswrapper[4796]: I0930 16:30:53.769327 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8a76e9ab-55ec-4394-94b3-668b7b146c81","Type":"ContainerStarted","Data":"b04152707f925534b2a04f0c82b5b6ac4d9febf97cb12cf53faef63e23ed4e81"} Sep 30 16:30:54 crc kubenswrapper[4796]: I0930 16:30:54.751310 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="720f2613-b6eb-43df-95a2-d05f980bb0c0" path="/var/lib/kubelet/pods/720f2613-b6eb-43df-95a2-d05f980bb0c0/volumes" Sep 30 16:30:54 crc kubenswrapper[4796]: I0930 16:30:54.780124 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8a76e9ab-55ec-4394-94b3-668b7b146c81","Type":"ContainerStarted","Data":"03a66edfd70d0f2352c07cb67fe4c9d2d8ad575ef8ec50e8e5cc511d71a6e010"} Sep 30 16:30:54 crc kubenswrapper[4796]: I0930 16:30:54.823145 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.823119283 podStartE2EDuration="2.823119283s" podCreationTimestamp="2025-09-30 16:30:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:54.805658222 +0000 UTC m=+1146.818936749" watchObservedRunningTime="2025-09-30 16:30:54.823119283 +0000 UTC m=+1146.836397820" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.245326 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.245397 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.245819 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.245871 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.249331 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.249392 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.465739 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-rgqtc"] Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.468965 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.483090 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-rgqtc"] Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.629498 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.629570 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.629612 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.629652 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n2p6\" (UniqueName: \"kubernetes.io/projected/02ef0f96-7297-4ba8-a571-0289dc0b4e07-kube-api-access-9n2p6\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.629688 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-config\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.629738 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.731398 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.731477 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n2p6\" (UniqueName: \"kubernetes.io/projected/02ef0f96-7297-4ba8-a571-0289dc0b4e07-kube-api-access-9n2p6\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.731521 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-config\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.731595 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.731654 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.731700 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.732400 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.732620 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.732964 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.733185 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-config\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.733223 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.751408 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n2p6\" (UniqueName: \"kubernetes.io/projected/02ef0f96-7297-4ba8-a571-0289dc0b4e07-kube-api-access-9n2p6\") pod \"dnsmasq-dns-59cf4bdb65-rgqtc\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:55 crc kubenswrapper[4796]: I0930 16:30:55.792911 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:56 crc kubenswrapper[4796]: I0930 16:30:56.252369 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-rgqtc"] Sep 30 16:30:56 crc kubenswrapper[4796]: W0930 16:30:56.253648 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02ef0f96_7297_4ba8_a571_0289dc0b4e07.slice/crio-90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d WatchSource:0}: Error finding container 90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d: Status 404 returned error can't find the container with id 90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d Sep 30 16:30:56 crc kubenswrapper[4796]: I0930 16:30:56.817599 4796 generic.go:334] "Generic (PLEG): container finished" podID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerID="988ba621f63a1e38a73f1c81041a63a6b256cc28ce5cc5466f8ba555d09e7697" exitCode=0 Sep 30 16:30:56 crc kubenswrapper[4796]: I0930 16:30:56.819370 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" event={"ID":"02ef0f96-7297-4ba8-a571-0289dc0b4e07","Type":"ContainerDied","Data":"988ba621f63a1e38a73f1c81041a63a6b256cc28ce5cc5466f8ba555d09e7697"} Sep 30 16:30:56 crc kubenswrapper[4796]: I0930 16:30:56.819472 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" event={"ID":"02ef0f96-7297-4ba8-a571-0289dc0b4e07","Type":"ContainerStarted","Data":"90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d"} Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.634291 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.634941 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-central-agent" containerID="cri-o://c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632" gracePeriod=30 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.635019 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-notification-agent" containerID="cri-o://c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45" gracePeriod=30 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.635084 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="sg-core" containerID="cri-o://64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777" gracePeriod=30 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.635087 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="proxy-httpd" containerID="cri-o://6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3" gracePeriod=30 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.739014 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.197:3000/\": read tcp 10.217.0.2:43942->10.217.0.197:3000: read: connection reset by peer" Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.779033 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.829687 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" event={"ID":"02ef0f96-7297-4ba8-a571-0289dc0b4e07","Type":"ContainerStarted","Data":"ad143d078d8ed653b598e264505cbf207b688cc9c70df7af71005abef2d4d7ca"} Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.830084 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.832720 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerID="64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777" exitCode=2 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.832892 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-log" containerID="cri-o://e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f" gracePeriod=30 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.833131 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerDied","Data":"64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777"} Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.833193 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-api" containerID="cri-o://9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2" gracePeriod=30 Sep 30 16:30:57 crc kubenswrapper[4796]: I0930 16:30:57.859163 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" podStartSLOduration=2.85914813 podStartE2EDuration="2.85914813s" podCreationTimestamp="2025-09-30 16:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:30:57.849542789 +0000 UTC m=+1149.862821326" watchObservedRunningTime="2025-09-30 16:30:57.85914813 +0000 UTC m=+1149.872426657" Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.169013 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.848084 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerID="6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3" exitCode=0 Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.848411 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerID="c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632" exitCode=0 Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.848184 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerDied","Data":"6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3"} Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.848486 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerDied","Data":"c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632"} Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.852271 4796 generic.go:334] "Generic (PLEG): container finished" podID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerID="e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f" exitCode=143 Sep 30 16:30:58 crc kubenswrapper[4796]: I0930 16:30:58.853515 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ee982da-2a4e-4e01-8315-acf0dd590839","Type":"ContainerDied","Data":"e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f"} Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.511440 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.619677 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zszsv\" (UniqueName: \"kubernetes.io/projected/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-kube-api-access-zszsv\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.619797 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-ceilometer-tls-certs\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.619838 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-config-data\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.619859 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-sg-core-conf-yaml\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620010 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-scripts\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620070 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-log-httpd\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620110 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-combined-ca-bundle\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620191 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-run-httpd\") pod \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\" (UID: \"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e\") " Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620560 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620828 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.620941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.625434 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-scripts" (OuterVolumeSpecName: "scripts") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.625471 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-kube-api-access-zszsv" (OuterVolumeSpecName: "kube-api-access-zszsv") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "kube-api-access-zszsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.651177 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.680160 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.718135 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-config-data" (OuterVolumeSpecName: "config-data") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.722878 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.722929 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.722948 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zszsv\" (UniqueName: \"kubernetes.io/projected/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-kube-api-access-zszsv\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.722966 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.723004 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.723021 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.742111 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" (UID: "8a531954-1e5a-4a38-a1fc-83d18e8e9e8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.825647 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.863415 4796 generic.go:334] "Generic (PLEG): container finished" podID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerID="c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45" exitCode=0 Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.863498 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerDied","Data":"c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45"} Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.863528 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a531954-1e5a-4a38-a1fc-83d18e8e9e8e","Type":"ContainerDied","Data":"e1b40130d6a6a9c14cef0d47643b2c356c8f05fa76d3adad9c7fcbd289163ca9"} Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.863549 4796 scope.go:117] "RemoveContainer" containerID="6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.863686 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.902131 4796 scope.go:117] "RemoveContainer" containerID="64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.918590 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.934421 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.944478 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:59 crc kubenswrapper[4796]: E0930 16:30:59.945166 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="proxy-httpd" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945190 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="proxy-httpd" Sep 30 16:30:59 crc kubenswrapper[4796]: E0930 16:30:59.945208 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-central-agent" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945219 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-central-agent" Sep 30 16:30:59 crc kubenswrapper[4796]: E0930 16:30:59.945245 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="sg-core" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945256 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="sg-core" Sep 30 16:30:59 crc kubenswrapper[4796]: E0930 16:30:59.945276 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-notification-agent" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945287 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-notification-agent" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945584 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-notification-agent" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945612 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="sg-core" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945631 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="proxy-httpd" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.945650 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" containerName="ceilometer-central-agent" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.948336 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.948591 4796 scope.go:117] "RemoveContainer" containerID="c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.951035 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.951035 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.951360 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.951649 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:30:59 crc kubenswrapper[4796]: I0930 16:30:59.984146 4796 scope.go:117] "RemoveContainer" containerID="c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.002672 4796 scope.go:117] "RemoveContainer" containerID="6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3" Sep 30 16:31:00 crc kubenswrapper[4796]: E0930 16:31:00.003084 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3\": container with ID starting with 6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3 not found: ID does not exist" containerID="6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.003119 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3"} err="failed to get container status \"6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3\": rpc error: code = NotFound desc = could not find container \"6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3\": container with ID starting with 6da0b98aed2cc72a232fc86a731d87fb3a6d138cefef4917f086b27069bd12c3 not found: ID does not exist" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.003148 4796 scope.go:117] "RemoveContainer" containerID="64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777" Sep 30 16:31:00 crc kubenswrapper[4796]: E0930 16:31:00.003424 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777\": container with ID starting with 64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777 not found: ID does not exist" containerID="64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.003457 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777"} err="failed to get container status \"64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777\": rpc error: code = NotFound desc = could not find container \"64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777\": container with ID starting with 64dfb01ec7bbc0d5ec7bbf436f205fe4b863aaf7a201b26e148217c57c21c777 not found: ID does not exist" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.003512 4796 scope.go:117] "RemoveContainer" containerID="c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45" Sep 30 16:31:00 crc kubenswrapper[4796]: E0930 16:31:00.003792 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45\": container with ID starting with c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45 not found: ID does not exist" containerID="c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.003835 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45"} err="failed to get container status \"c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45\": rpc error: code = NotFound desc = could not find container \"c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45\": container with ID starting with c785cbff76f10f85a7ddb7092e22742c59b636c094ddd69d8b6b81f31ac9fb45 not found: ID does not exist" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.003872 4796 scope.go:117] "RemoveContainer" containerID="c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632" Sep 30 16:31:00 crc kubenswrapper[4796]: E0930 16:31:00.004915 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632\": container with ID starting with c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632 not found: ID does not exist" containerID="c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.004949 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632"} err="failed to get container status \"c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632\": rpc error: code = NotFound desc = could not find container \"c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632\": container with ID starting with c3fe91be56e97b063be5d08998a84306ffd51dd7a3af0e0c9b052bdef5f32632 not found: ID does not exist" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.132772 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-config-data\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.132830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.132881 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-run-httpd\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.132928 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.132992 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6ktr\" (UniqueName: \"kubernetes.io/projected/d67076ba-6fad-4096-8aed-c76ad237a19a-kube-api-access-g6ktr\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.133017 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-scripts\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.133040 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-log-httpd\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.133110 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234544 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234600 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-run-httpd\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234637 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234686 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6ktr\" (UniqueName: \"kubernetes.io/projected/d67076ba-6fad-4096-8aed-c76ad237a19a-kube-api-access-g6ktr\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234713 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-scripts\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234728 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-log-httpd\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234761 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.234815 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-config-data\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.236121 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-run-httpd\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.236552 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-log-httpd\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.239682 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.240303 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-config-data\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.241399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.241948 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.243524 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-scripts\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.257619 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6ktr\" (UniqueName: \"kubernetes.io/projected/d67076ba-6fad-4096-8aed-c76ad237a19a-kube-api-access-g6ktr\") pod \"ceilometer-0\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.279696 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:31:00 crc kubenswrapper[4796]: W0930 16:31:00.742867 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd67076ba_6fad_4096_8aed_c76ad237a19a.slice/crio-870540ea0b26857d8bf980a47dee42871a19bb19d798f9551f731e5919be3af3 WatchSource:0}: Error finding container 870540ea0b26857d8bf980a47dee42871a19bb19d798f9551f731e5919be3af3: Status 404 returned error can't find the container with id 870540ea0b26857d8bf980a47dee42871a19bb19d798f9551f731e5919be3af3 Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.748633 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.755840 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a531954-1e5a-4a38-a1fc-83d18e8e9e8e" path="/var/lib/kubelet/pods/8a531954-1e5a-4a38-a1fc-83d18e8e9e8e/volumes" Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.757010 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:00 crc kubenswrapper[4796]: I0930 16:31:00.873544 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerStarted","Data":"870540ea0b26857d8bf980a47dee42871a19bb19d798f9551f731e5919be3af3"} Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.419999 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.559426 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-combined-ca-bundle\") pod \"4ee982da-2a4e-4e01-8315-acf0dd590839\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.560715 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-config-data\") pod \"4ee982da-2a4e-4e01-8315-acf0dd590839\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.560768 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pws5\" (UniqueName: \"kubernetes.io/projected/4ee982da-2a4e-4e01-8315-acf0dd590839-kube-api-access-7pws5\") pod \"4ee982da-2a4e-4e01-8315-acf0dd590839\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.560801 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee982da-2a4e-4e01-8315-acf0dd590839-logs\") pod \"4ee982da-2a4e-4e01-8315-acf0dd590839\" (UID: \"4ee982da-2a4e-4e01-8315-acf0dd590839\") " Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.561371 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ee982da-2a4e-4e01-8315-acf0dd590839-logs" (OuterVolumeSpecName: "logs") pod "4ee982da-2a4e-4e01-8315-acf0dd590839" (UID: "4ee982da-2a4e-4e01-8315-acf0dd590839"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.564084 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ee982da-2a4e-4e01-8315-acf0dd590839-kube-api-access-7pws5" (OuterVolumeSpecName: "kube-api-access-7pws5") pod "4ee982da-2a4e-4e01-8315-acf0dd590839" (UID: "4ee982da-2a4e-4e01-8315-acf0dd590839"). InnerVolumeSpecName "kube-api-access-7pws5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.590791 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-config-data" (OuterVolumeSpecName: "config-data") pod "4ee982da-2a4e-4e01-8315-acf0dd590839" (UID: "4ee982da-2a4e-4e01-8315-acf0dd590839"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.594160 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ee982da-2a4e-4e01-8315-acf0dd590839" (UID: "4ee982da-2a4e-4e01-8315-acf0dd590839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.663263 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.663294 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pws5\" (UniqueName: \"kubernetes.io/projected/4ee982da-2a4e-4e01-8315-acf0dd590839-kube-api-access-7pws5\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.663308 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee982da-2a4e-4e01-8315-acf0dd590839-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.663318 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee982da-2a4e-4e01-8315-acf0dd590839-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.886662 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerStarted","Data":"f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f"} Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.888677 4796 generic.go:334] "Generic (PLEG): container finished" podID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerID="9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2" exitCode=0 Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.888708 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ee982da-2a4e-4e01-8315-acf0dd590839","Type":"ContainerDied","Data":"9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2"} Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.888723 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ee982da-2a4e-4e01-8315-acf0dd590839","Type":"ContainerDied","Data":"5fe808b61b62eee5bb37314ab7690635124f5ec060a2235d50ebd44adf9343c8"} Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.888741 4796 scope.go:117] "RemoveContainer" containerID="9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.888820 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.938134 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.938848 4796 scope.go:117] "RemoveContainer" containerID="e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.957421 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.971049 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:01 crc kubenswrapper[4796]: E0930 16:31:01.971541 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-api" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.971563 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-api" Sep 30 16:31:01 crc kubenswrapper[4796]: E0930 16:31:01.971588 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-log" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.971597 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-log" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.971825 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-log" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.971859 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" containerName="nova-api-api" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.973167 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.975599 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.975760 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.975878 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.986215 4796 scope.go:117] "RemoveContainer" containerID="9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.990553 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:01 crc kubenswrapper[4796]: E0930 16:31:01.993339 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2\": container with ID starting with 9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2 not found: ID does not exist" containerID="9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.993387 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2"} err="failed to get container status \"9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2\": rpc error: code = NotFound desc = could not find container \"9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2\": container with ID starting with 9ecd08192ccdadfbe86d095a6a58af230a1d034d3c2b04b019448c6fdb58d1c2 not found: ID does not exist" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.993420 4796 scope.go:117] "RemoveContainer" containerID="e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f" Sep 30 16:31:01 crc kubenswrapper[4796]: E0930 16:31:01.994466 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f\": container with ID starting with e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f not found: ID does not exist" containerID="e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f" Sep 30 16:31:01 crc kubenswrapper[4796]: I0930 16:31:01.994509 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f"} err="failed to get container status \"e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f\": rpc error: code = NotFound desc = could not find container \"e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f\": container with ID starting with e997f51a2736ca87f73a78a5ee6c616b250820fc8ea132a7e5a961afe920ed1f not found: ID does not exist" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.071482 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-config-data\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.071552 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.071585 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cac8c42-838b-49bc-8199-1dfb8c87de38-logs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.071608 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-public-tls-certs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.071855 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.072134 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l5r2\" (UniqueName: \"kubernetes.io/projected/5cac8c42-838b-49bc-8199-1dfb8c87de38-kube-api-access-2l5r2\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.174183 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-config-data\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.174282 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.174335 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cac8c42-838b-49bc-8199-1dfb8c87de38-logs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.174360 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-public-tls-certs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.174394 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.174444 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l5r2\" (UniqueName: \"kubernetes.io/projected/5cac8c42-838b-49bc-8199-1dfb8c87de38-kube-api-access-2l5r2\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.175381 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cac8c42-838b-49bc-8199-1dfb8c87de38-logs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.179194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.179327 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.179683 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-public-tls-certs\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.182568 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-config-data\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.190444 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l5r2\" (UniqueName: \"kubernetes.io/projected/5cac8c42-838b-49bc-8199-1dfb8c87de38-kube-api-access-2l5r2\") pod \"nova-api-0\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.300531 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.752084 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ee982da-2a4e-4e01-8315-acf0dd590839" path="/var/lib/kubelet/pods/4ee982da-2a4e-4e01-8315-acf0dd590839/volumes" Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.792359 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.905282 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerStarted","Data":"84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953"} Sep 30 16:31:02 crc kubenswrapper[4796]: I0930 16:31:02.906316 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5cac8c42-838b-49bc-8199-1dfb8c87de38","Type":"ContainerStarted","Data":"adc271ffda80bc4e5f27c740f44682238de40c54283688f64fc66eaf911664cf"} Sep 30 16:31:03 crc kubenswrapper[4796]: I0930 16:31:03.168923 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:31:03 crc kubenswrapper[4796]: I0930 16:31:03.186536 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:31:03 crc kubenswrapper[4796]: I0930 16:31:03.934780 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5cac8c42-838b-49bc-8199-1dfb8c87de38","Type":"ContainerStarted","Data":"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d"} Sep 30 16:31:03 crc kubenswrapper[4796]: I0930 16:31:03.935109 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5cac8c42-838b-49bc-8199-1dfb8c87de38","Type":"ContainerStarted","Data":"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d"} Sep 30 16:31:03 crc kubenswrapper[4796]: I0930 16:31:03.953741 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 30 16:31:03 crc kubenswrapper[4796]: I0930 16:31:03.970443 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.97041927 podStartE2EDuration="2.97041927s" podCreationTimestamp="2025-09-30 16:31:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:31:03.961393726 +0000 UTC m=+1155.974672263" watchObservedRunningTime="2025-09-30 16:31:03.97041927 +0000 UTC m=+1155.983697807" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.160254 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-q8zkj"] Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.161391 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.165247 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.165567 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.172646 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q8zkj"] Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.314523 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-scripts\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.314578 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-config-data\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.314698 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g55w\" (UniqueName: \"kubernetes.io/projected/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-kube-api-access-8g55w\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.314774 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.416593 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g55w\" (UniqueName: \"kubernetes.io/projected/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-kube-api-access-8g55w\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.416662 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.416720 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-scripts\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.416745 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-config-data\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.425511 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.429822 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-scripts\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.438567 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g55w\" (UniqueName: \"kubernetes.io/projected/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-kube-api-access-8g55w\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.439094 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-config-data\") pod \"nova-cell1-cell-mapping-q8zkj\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.489723 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.946671 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerStarted","Data":"1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4"} Sep 30 16:31:04 crc kubenswrapper[4796]: I0930 16:31:04.988268 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q8zkj"] Sep 30 16:31:04 crc kubenswrapper[4796]: W0930 16:31:04.999573 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6da3535_7e08_4b79_a4d5_6c94ab7ff0a4.slice/crio-66a3e3d70f3ee6e2c1e9a521da887cfb7320b58c4282dd897df15d7197c1bba9 WatchSource:0}: Error finding container 66a3e3d70f3ee6e2c1e9a521da887cfb7320b58c4282dd897df15d7197c1bba9: Status 404 returned error can't find the container with id 66a3e3d70f3ee6e2c1e9a521da887cfb7320b58c4282dd897df15d7197c1bba9 Sep 30 16:31:05 crc kubenswrapper[4796]: I0930 16:31:05.794447 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:31:05 crc kubenswrapper[4796]: I0930 16:31:05.897399 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hn7wq"] Sep 30 16:31:05 crc kubenswrapper[4796]: I0930 16:31:05.897879 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerName="dnsmasq-dns" containerID="cri-o://e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6" gracePeriod=10 Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.002140 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q8zkj" event={"ID":"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4","Type":"ContainerStarted","Data":"ac54fdb717ec98f5bc5aa9bede4e1e7f76dd72191db4ea4639250c35c4d24508"} Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.002186 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q8zkj" event={"ID":"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4","Type":"ContainerStarted","Data":"66a3e3d70f3ee6e2c1e9a521da887cfb7320b58c4282dd897df15d7197c1bba9"} Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.050496 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-q8zkj" podStartSLOduration=2.050476449 podStartE2EDuration="2.050476449s" podCreationTimestamp="2025-09-30 16:31:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:31:06.042261059 +0000 UTC m=+1158.055539586" watchObservedRunningTime="2025-09-30 16:31:06.050476449 +0000 UTC m=+1158.063754976" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.539165 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.677121 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-swift-storage-0\") pod \"c3e1c88b-b3be-443d-a971-7433e3d868e7\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.677481 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-config\") pod \"c3e1c88b-b3be-443d-a971-7433e3d868e7\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.677510 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9m69\" (UniqueName: \"kubernetes.io/projected/c3e1c88b-b3be-443d-a971-7433e3d868e7-kube-api-access-p9m69\") pod \"c3e1c88b-b3be-443d-a971-7433e3d868e7\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.677589 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-nb\") pod \"c3e1c88b-b3be-443d-a971-7433e3d868e7\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.677675 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-sb\") pod \"c3e1c88b-b3be-443d-a971-7433e3d868e7\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.677796 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-svc\") pod \"c3e1c88b-b3be-443d-a971-7433e3d868e7\" (UID: \"c3e1c88b-b3be-443d-a971-7433e3d868e7\") " Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.684279 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3e1c88b-b3be-443d-a971-7433e3d868e7-kube-api-access-p9m69" (OuterVolumeSpecName: "kube-api-access-p9m69") pod "c3e1c88b-b3be-443d-a971-7433e3d868e7" (UID: "c3e1c88b-b3be-443d-a971-7433e3d868e7"). InnerVolumeSpecName "kube-api-access-p9m69". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.730683 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c3e1c88b-b3be-443d-a971-7433e3d868e7" (UID: "c3e1c88b-b3be-443d-a971-7433e3d868e7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.736587 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c3e1c88b-b3be-443d-a971-7433e3d868e7" (UID: "c3e1c88b-b3be-443d-a971-7433e3d868e7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.740173 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-config" (OuterVolumeSpecName: "config") pod "c3e1c88b-b3be-443d-a971-7433e3d868e7" (UID: "c3e1c88b-b3be-443d-a971-7433e3d868e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.742686 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3e1c88b-b3be-443d-a971-7433e3d868e7" (UID: "c3e1c88b-b3be-443d-a971-7433e3d868e7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.744465 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c3e1c88b-b3be-443d-a971-7433e3d868e7" (UID: "c3e1c88b-b3be-443d-a971-7433e3d868e7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.780774 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.780812 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.780828 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.780860 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.780876 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e1c88b-b3be-443d-a971-7433e3d868e7-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:06 crc kubenswrapper[4796]: I0930 16:31:06.780886 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9m69\" (UniqueName: \"kubernetes.io/projected/c3e1c88b-b3be-443d-a971-7433e3d868e7-kube-api-access-p9m69\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.014116 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerStarted","Data":"17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c"} Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.015292 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.017591 4796 generic.go:334] "Generic (PLEG): container finished" podID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerID="e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6" exitCode=0 Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.018482 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.019089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" event={"ID":"c3e1c88b-b3be-443d-a971-7433e3d868e7","Type":"ContainerDied","Data":"e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6"} Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.019164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hn7wq" event={"ID":"c3e1c88b-b3be-443d-a971-7433e3d868e7","Type":"ContainerDied","Data":"34ceb3283afbf24ee1049b216c0e82d3711dfc94528092a7150e0b2f177bf621"} Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.019189 4796 scope.go:117] "RemoveContainer" containerID="e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.041031 4796 scope.go:117] "RemoveContainer" containerID="151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.062126 4796 scope.go:117] "RemoveContainer" containerID="e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6" Sep 30 16:31:07 crc kubenswrapper[4796]: E0930 16:31:07.062473 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6\": container with ID starting with e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6 not found: ID does not exist" containerID="e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.062507 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6"} err="failed to get container status \"e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6\": rpc error: code = NotFound desc = could not find container \"e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6\": container with ID starting with e611aef29d9ca6da05f33287fdc069e5649a5de9c994ed0054e4284abe2719d6 not found: ID does not exist" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.062529 4796 scope.go:117] "RemoveContainer" containerID="151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9" Sep 30 16:31:07 crc kubenswrapper[4796]: E0930 16:31:07.062910 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9\": container with ID starting with 151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9 not found: ID does not exist" containerID="151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.062928 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9"} err="failed to get container status \"151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9\": rpc error: code = NotFound desc = could not find container \"151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9\": container with ID starting with 151e9ea382852fc5403f35cf7465b612933c651c9f8cdfe0380404922c4ea4b9 not found: ID does not exist" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.065198 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.957003704 podStartE2EDuration="8.065182087s" podCreationTimestamp="2025-09-30 16:30:59 +0000 UTC" firstStartedPulling="2025-09-30 16:31:00.748334339 +0000 UTC m=+1152.761612866" lastFinishedPulling="2025-09-30 16:31:05.856512722 +0000 UTC m=+1157.869791249" observedRunningTime="2025-09-30 16:31:07.041223776 +0000 UTC m=+1159.054502293" watchObservedRunningTime="2025-09-30 16:31:07.065182087 +0000 UTC m=+1159.078460614" Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.070918 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hn7wq"] Sep 30 16:31:07 crc kubenswrapper[4796]: I0930 16:31:07.082291 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hn7wq"] Sep 30 16:31:08 crc kubenswrapper[4796]: I0930 16:31:08.755374 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" path="/var/lib/kubelet/pods/c3e1c88b-b3be-443d-a971-7433e3d868e7/volumes" Sep 30 16:31:10 crc kubenswrapper[4796]: I0930 16:31:10.065797 4796 generic.go:334] "Generic (PLEG): container finished" podID="c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" containerID="ac54fdb717ec98f5bc5aa9bede4e1e7f76dd72191db4ea4639250c35c4d24508" exitCode=0 Sep 30 16:31:10 crc kubenswrapper[4796]: I0930 16:31:10.065888 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q8zkj" event={"ID":"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4","Type":"ContainerDied","Data":"ac54fdb717ec98f5bc5aa9bede4e1e7f76dd72191db4ea4639250c35c4d24508"} Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.627465 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.785537 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-combined-ca-bundle\") pod \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.785690 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g55w\" (UniqueName: \"kubernetes.io/projected/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-kube-api-access-8g55w\") pod \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.785770 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-config-data\") pod \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.785916 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-scripts\") pod \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\" (UID: \"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4\") " Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.790673 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-scripts" (OuterVolumeSpecName: "scripts") pod "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" (UID: "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.800191 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-kube-api-access-8g55w" (OuterVolumeSpecName: "kube-api-access-8g55w") pod "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" (UID: "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4"). InnerVolumeSpecName "kube-api-access-8g55w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.812912 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-config-data" (OuterVolumeSpecName: "config-data") pod "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" (UID: "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.820033 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" (UID: "c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.896697 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.896921 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.897097 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:11 crc kubenswrapper[4796]: I0930 16:31:11.897217 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g55w\" (UniqueName: \"kubernetes.io/projected/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4-kube-api-access-8g55w\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.088943 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q8zkj" event={"ID":"c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4","Type":"ContainerDied","Data":"66a3e3d70f3ee6e2c1e9a521da887cfb7320b58c4282dd897df15d7197c1bba9"} Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.088998 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66a3e3d70f3ee6e2c1e9a521da887cfb7320b58c4282dd897df15d7197c1bba9" Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.089013 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q8zkj" Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.295782 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.296120 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-log" containerID="cri-o://72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d" gracePeriod=30 Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.296429 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-api" containerID="cri-o://31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d" gracePeriod=30 Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.322775 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.325121 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0d226c88-71f5-4831-8576-4c8d02c0fb30" containerName="nova-scheduler-scheduler" containerID="cri-o://f5d2055e0daa8a6c32500446e823ae401143de578c663dc0332c12e79c19b0ae" gracePeriod=30 Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.413442 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.414127 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-log" containerID="cri-o://8923fde2eba2d93d4cf942b68a5fadeec1b42fcb138fe20b5d3238a83dffc15f" gracePeriod=30 Sep 30 16:31:12 crc kubenswrapper[4796]: I0930 16:31:12.414578 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-metadata" containerID="cri-o://9e0c4fa9f51e60247a9f5b85ab6550c10ea58c4fae85253c842c7f6f8634a34f" gracePeriod=30 Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.103182 4796 generic.go:334] "Generic (PLEG): container finished" podID="7e16f45a-71c1-411e-9084-c950519e768d" containerID="8923fde2eba2d93d4cf942b68a5fadeec1b42fcb138fe20b5d3238a83dffc15f" exitCode=143 Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.103467 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e16f45a-71c1-411e-9084-c950519e768d","Type":"ContainerDied","Data":"8923fde2eba2d93d4cf942b68a5fadeec1b42fcb138fe20b5d3238a83dffc15f"} Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.105878 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.105932 4796 generic.go:334] "Generic (PLEG): container finished" podID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerID="31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d" exitCode=0 Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.105989 4796 generic.go:334] "Generic (PLEG): container finished" podID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerID="72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d" exitCode=143 Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.106047 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5cac8c42-838b-49bc-8199-1dfb8c87de38","Type":"ContainerDied","Data":"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d"} Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.106080 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5cac8c42-838b-49bc-8199-1dfb8c87de38","Type":"ContainerDied","Data":"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d"} Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.106118 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5cac8c42-838b-49bc-8199-1dfb8c87de38","Type":"ContainerDied","Data":"adc271ffda80bc4e5f27c740f44682238de40c54283688f64fc66eaf911664cf"} Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.106137 4796 scope.go:117] "RemoveContainer" containerID="31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.108293 4796 generic.go:334] "Generic (PLEG): container finished" podID="0d226c88-71f5-4831-8576-4c8d02c0fb30" containerID="f5d2055e0daa8a6c32500446e823ae401143de578c663dc0332c12e79c19b0ae" exitCode=0 Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.108358 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d226c88-71f5-4831-8576-4c8d02c0fb30","Type":"ContainerDied","Data":"f5d2055e0daa8a6c32500446e823ae401143de578c663dc0332c12e79c19b0ae"} Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.141399 4796 scope.go:117] "RemoveContainer" containerID="72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.170613 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.181127 4796 scope.go:117] "RemoveContainer" containerID="31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d" Sep 30 16:31:13 crc kubenswrapper[4796]: E0930 16:31:13.181544 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d\": container with ID starting with 31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d not found: ID does not exist" containerID="31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.181598 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d"} err="failed to get container status \"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d\": rpc error: code = NotFound desc = could not find container \"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d\": container with ID starting with 31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d not found: ID does not exist" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.181644 4796 scope.go:117] "RemoveContainer" containerID="72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d" Sep 30 16:31:13 crc kubenswrapper[4796]: E0930 16:31:13.182318 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d\": container with ID starting with 72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d not found: ID does not exist" containerID="72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.182541 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d"} err="failed to get container status \"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d\": rpc error: code = NotFound desc = could not find container \"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d\": container with ID starting with 72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d not found: ID does not exist" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.182573 4796 scope.go:117] "RemoveContainer" containerID="31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.184084 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d"} err="failed to get container status \"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d\": rpc error: code = NotFound desc = could not find container \"31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d\": container with ID starting with 31bba8eb77521ab76ec0e77b0387c0944525c93c52876ab803fe61983ed3079d not found: ID does not exist" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.184112 4796 scope.go:117] "RemoveContainer" containerID="72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.184826 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d"} err="failed to get container status \"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d\": rpc error: code = NotFound desc = could not find container \"72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d\": container with ID starting with 72eb1e49a1bcf736efc8eae618b2f0ba914c3c6012dc1f2759e9e28e39a58a7d not found: ID does not exist" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.224590 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-public-tls-certs\") pod \"5cac8c42-838b-49bc-8199-1dfb8c87de38\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.224739 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cac8c42-838b-49bc-8199-1dfb8c87de38-logs\") pod \"5cac8c42-838b-49bc-8199-1dfb8c87de38\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.224812 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-combined-ca-bundle\") pod \"0d226c88-71f5-4831-8576-4c8d02c0fb30\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.224869 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l5r2\" (UniqueName: \"kubernetes.io/projected/5cac8c42-838b-49bc-8199-1dfb8c87de38-kube-api-access-2l5r2\") pod \"5cac8c42-838b-49bc-8199-1dfb8c87de38\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.224997 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-config-data\") pod \"0d226c88-71f5-4831-8576-4c8d02c0fb30\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.225129 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5jhf\" (UniqueName: \"kubernetes.io/projected/0d226c88-71f5-4831-8576-4c8d02c0fb30-kube-api-access-f5jhf\") pod \"0d226c88-71f5-4831-8576-4c8d02c0fb30\" (UID: \"0d226c88-71f5-4831-8576-4c8d02c0fb30\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.225170 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-combined-ca-bundle\") pod \"5cac8c42-838b-49bc-8199-1dfb8c87de38\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.225217 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-internal-tls-certs\") pod \"5cac8c42-838b-49bc-8199-1dfb8c87de38\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.225299 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-config-data\") pod \"5cac8c42-838b-49bc-8199-1dfb8c87de38\" (UID: \"5cac8c42-838b-49bc-8199-1dfb8c87de38\") " Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.225472 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cac8c42-838b-49bc-8199-1dfb8c87de38-logs" (OuterVolumeSpecName: "logs") pod "5cac8c42-838b-49bc-8199-1dfb8c87de38" (UID: "5cac8c42-838b-49bc-8199-1dfb8c87de38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.226130 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cac8c42-838b-49bc-8199-1dfb8c87de38-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.234910 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cac8c42-838b-49bc-8199-1dfb8c87de38-kube-api-access-2l5r2" (OuterVolumeSpecName: "kube-api-access-2l5r2") pod "5cac8c42-838b-49bc-8199-1dfb8c87de38" (UID: "5cac8c42-838b-49bc-8199-1dfb8c87de38"). InnerVolumeSpecName "kube-api-access-2l5r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.234951 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d226c88-71f5-4831-8576-4c8d02c0fb30-kube-api-access-f5jhf" (OuterVolumeSpecName: "kube-api-access-f5jhf") pod "0d226c88-71f5-4831-8576-4c8d02c0fb30" (UID: "0d226c88-71f5-4831-8576-4c8d02c0fb30"). InnerVolumeSpecName "kube-api-access-f5jhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.256724 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d226c88-71f5-4831-8576-4c8d02c0fb30" (UID: "0d226c88-71f5-4831-8576-4c8d02c0fb30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.266408 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cac8c42-838b-49bc-8199-1dfb8c87de38" (UID: "5cac8c42-838b-49bc-8199-1dfb8c87de38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.276434 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-config-data" (OuterVolumeSpecName: "config-data") pod "0d226c88-71f5-4831-8576-4c8d02c0fb30" (UID: "0d226c88-71f5-4831-8576-4c8d02c0fb30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.278170 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-config-data" (OuterVolumeSpecName: "config-data") pod "5cac8c42-838b-49bc-8199-1dfb8c87de38" (UID: "5cac8c42-838b-49bc-8199-1dfb8c87de38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.310663 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5cac8c42-838b-49bc-8199-1dfb8c87de38" (UID: "5cac8c42-838b-49bc-8199-1dfb8c87de38"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.311472 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5cac8c42-838b-49bc-8199-1dfb8c87de38" (UID: "5cac8c42-838b-49bc-8199-1dfb8c87de38"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328446 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328475 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l5r2\" (UniqueName: \"kubernetes.io/projected/5cac8c42-838b-49bc-8199-1dfb8c87de38-kube-api-access-2l5r2\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328488 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d226c88-71f5-4831-8576-4c8d02c0fb30-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328526 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5jhf\" (UniqueName: \"kubernetes.io/projected/0d226c88-71f5-4831-8576-4c8d02c0fb30-kube-api-access-f5jhf\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328536 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328565 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328575 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:13 crc kubenswrapper[4796]: I0930 16:31:13.328591 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cac8c42-838b-49bc-8199-1dfb8c87de38-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.120622 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.123865 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d226c88-71f5-4831-8576-4c8d02c0fb30","Type":"ContainerDied","Data":"db691c3867efd2379c37fc22fb1c8f3deda15a55119cadd9193fd22e12f88984"} Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.123918 4796 scope.go:117] "RemoveContainer" containerID="f5d2055e0daa8a6c32500446e823ae401143de578c663dc0332c12e79c19b0ae" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.123939 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.164282 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.181663 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.193802 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.208227 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236318 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: E0930 16:31:14.236802 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-log" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236821 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-log" Sep 30 16:31:14 crc kubenswrapper[4796]: E0930 16:31:14.236846 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" containerName="nova-manage" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236855 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" containerName="nova-manage" Sep 30 16:31:14 crc kubenswrapper[4796]: E0930 16:31:14.236881 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerName="dnsmasq-dns" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236890 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerName="dnsmasq-dns" Sep 30 16:31:14 crc kubenswrapper[4796]: E0930 16:31:14.236908 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-api" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236918 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-api" Sep 30 16:31:14 crc kubenswrapper[4796]: E0930 16:31:14.236934 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d226c88-71f5-4831-8576-4c8d02c0fb30" containerName="nova-scheduler-scheduler" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236943 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d226c88-71f5-4831-8576-4c8d02c0fb30" containerName="nova-scheduler-scheduler" Sep 30 16:31:14 crc kubenswrapper[4796]: E0930 16:31:14.236960 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerName="init" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.236968 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerName="init" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.237218 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d226c88-71f5-4831-8576-4c8d02c0fb30" containerName="nova-scheduler-scheduler" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.237241 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-api" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.237258 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" containerName="nova-api-log" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.237280 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" containerName="nova-manage" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.237297 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3e1c88b-b3be-443d-a971-7433e3d868e7" containerName="dnsmasq-dns" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.238459 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.241490 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.241661 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.241762 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.247956 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.249390 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.251359 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.259539 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.270513 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.350872 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6185e29-465b-4ab1-b963-88ff87922e39-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.350946 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351003 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-config-data\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351029 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb4sj\" (UniqueName: \"kubernetes.io/projected/a6185e29-465b-4ab1-b963-88ff87922e39-kube-api-access-fb4sj\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351147 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351173 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-public-tls-certs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351205 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s7rt\" (UniqueName: \"kubernetes.io/projected/dec95fef-d221-4739-849e-5e071ce2311a-kube-api-access-2s7rt\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6185e29-465b-4ab1-b963-88ff87922e39-config-data\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.351304 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dec95fef-d221-4739-849e-5e071ce2311a-logs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453058 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453126 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb4sj\" (UniqueName: \"kubernetes.io/projected/a6185e29-465b-4ab1-b963-88ff87922e39-kube-api-access-fb4sj\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453155 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-config-data\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453224 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453247 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-public-tls-certs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453274 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s7rt\" (UniqueName: \"kubernetes.io/projected/dec95fef-d221-4739-849e-5e071ce2311a-kube-api-access-2s7rt\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453320 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6185e29-465b-4ab1-b963-88ff87922e39-config-data\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453366 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dec95fef-d221-4739-849e-5e071ce2311a-logs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.453427 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6185e29-465b-4ab1-b963-88ff87922e39-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.455003 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dec95fef-d221-4739-849e-5e071ce2311a-logs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.458819 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.459347 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6185e29-465b-4ab1-b963-88ff87922e39-config-data\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.459655 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.460854 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6185e29-465b-4ab1-b963-88ff87922e39-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.461857 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-public-tls-certs\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.469492 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dec95fef-d221-4739-849e-5e071ce2311a-config-data\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.478169 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb4sj\" (UniqueName: \"kubernetes.io/projected/a6185e29-465b-4ab1-b963-88ff87922e39-kube-api-access-fb4sj\") pod \"nova-scheduler-0\" (UID: \"a6185e29-465b-4ab1-b963-88ff87922e39\") " pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.480723 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s7rt\" (UniqueName: \"kubernetes.io/projected/dec95fef-d221-4739-849e-5e071ce2311a-kube-api-access-2s7rt\") pod \"nova-api-0\" (UID: \"dec95fef-d221-4739-849e-5e071ce2311a\") " pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.563569 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.576852 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.753119 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d226c88-71f5-4831-8576-4c8d02c0fb30" path="/var/lib/kubelet/pods/0d226c88-71f5-4831-8576-4c8d02c0fb30/volumes" Sep 30 16:31:14 crc kubenswrapper[4796]: I0930 16:31:14.754269 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cac8c42-838b-49bc-8199-1dfb8c87de38" path="/var/lib/kubelet/pods/5cac8c42-838b-49bc-8199-1dfb8c87de38/volumes" Sep 30 16:31:15 crc kubenswrapper[4796]: I0930 16:31:15.030187 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 16:31:15 crc kubenswrapper[4796]: W0930 16:31:15.034160 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddec95fef_d221_4739_849e_5e071ce2311a.slice/crio-6d35c06eecf0ddab5852004a7b3f8ce9cf663560967d46cca33ace888d2c9d72 WatchSource:0}: Error finding container 6d35c06eecf0ddab5852004a7b3f8ce9cf663560967d46cca33ace888d2c9d72: Status 404 returned error can't find the container with id 6d35c06eecf0ddab5852004a7b3f8ce9cf663560967d46cca33ace888d2c9d72 Sep 30 16:31:15 crc kubenswrapper[4796]: I0930 16:31:15.116459 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 16:31:15 crc kubenswrapper[4796]: W0930 16:31:15.131923 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6185e29_465b_4ab1_b963_88ff87922e39.slice/crio-e126a8ebcc7585a5b035344e882f9f5c96332c2b2a7e47c50e6355dd77899fac WatchSource:0}: Error finding container e126a8ebcc7585a5b035344e882f9f5c96332c2b2a7e47c50e6355dd77899fac: Status 404 returned error can't find the container with id e126a8ebcc7585a5b035344e882f9f5c96332c2b2a7e47c50e6355dd77899fac Sep 30 16:31:15 crc kubenswrapper[4796]: I0930 16:31:15.138601 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dec95fef-d221-4739-849e-5e071ce2311a","Type":"ContainerStarted","Data":"6d35c06eecf0ddab5852004a7b3f8ce9cf663560967d46cca33ace888d2c9d72"} Sep 30 16:31:15 crc kubenswrapper[4796]: I0930 16:31:15.925537 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:37230->10.217.0.198:8775: read: connection reset by peer" Sep 30 16:31:15 crc kubenswrapper[4796]: I0930 16:31:15.925537 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:37224->10.217.0.198:8775: read: connection reset by peer" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.152225 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a6185e29-465b-4ab1-b963-88ff87922e39","Type":"ContainerStarted","Data":"e4abf3dd7b71ed3fe19b6d859ac432ba41f998cf4d8dfeac557f96b1b937863f"} Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.152272 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a6185e29-465b-4ab1-b963-88ff87922e39","Type":"ContainerStarted","Data":"e126a8ebcc7585a5b035344e882f9f5c96332c2b2a7e47c50e6355dd77899fac"} Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.158013 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dec95fef-d221-4739-849e-5e071ce2311a","Type":"ContainerStarted","Data":"5ca88d58ed6f49bfeede6d864bfbaa61b1db00f82b6aaf24c7365ca5d064ab2e"} Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.158041 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dec95fef-d221-4739-849e-5e071ce2311a","Type":"ContainerStarted","Data":"f85f6e7e073367262a07944d1d0efa6356ddbdaf4901aae28fe8845718f2af53"} Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.163293 4796 generic.go:334] "Generic (PLEG): container finished" podID="7e16f45a-71c1-411e-9084-c950519e768d" containerID="9e0c4fa9f51e60247a9f5b85ab6550c10ea58c4fae85253c842c7f6f8634a34f" exitCode=0 Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.163348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e16f45a-71c1-411e-9084-c950519e768d","Type":"ContainerDied","Data":"9e0c4fa9f51e60247a9f5b85ab6550c10ea58c4fae85253c842c7f6f8634a34f"} Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.174967 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.174950931 podStartE2EDuration="2.174950931s" podCreationTimestamp="2025-09-30 16:31:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:31:16.169209564 +0000 UTC m=+1168.182488111" watchObservedRunningTime="2025-09-30 16:31:16.174950931 +0000 UTC m=+1168.188229458" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.194152 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.19413345 podStartE2EDuration="2.19413345s" podCreationTimestamp="2025-09-30 16:31:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:31:16.187239699 +0000 UTC m=+1168.200518226" watchObservedRunningTime="2025-09-30 16:31:16.19413345 +0000 UTC m=+1168.207411977" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.414644 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.493997 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-nova-metadata-tls-certs\") pod \"7e16f45a-71c1-411e-9084-c950519e768d\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.494185 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whb5h\" (UniqueName: \"kubernetes.io/projected/7e16f45a-71c1-411e-9084-c950519e768d-kube-api-access-whb5h\") pod \"7e16f45a-71c1-411e-9084-c950519e768d\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.494243 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-config-data\") pod \"7e16f45a-71c1-411e-9084-c950519e768d\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.494358 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-combined-ca-bundle\") pod \"7e16f45a-71c1-411e-9084-c950519e768d\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.494436 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e16f45a-71c1-411e-9084-c950519e768d-logs\") pod \"7e16f45a-71c1-411e-9084-c950519e768d\" (UID: \"7e16f45a-71c1-411e-9084-c950519e768d\") " Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.495474 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e16f45a-71c1-411e-9084-c950519e768d-logs" (OuterVolumeSpecName: "logs") pod "7e16f45a-71c1-411e-9084-c950519e768d" (UID: "7e16f45a-71c1-411e-9084-c950519e768d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.518339 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e16f45a-71c1-411e-9084-c950519e768d-kube-api-access-whb5h" (OuterVolumeSpecName: "kube-api-access-whb5h") pod "7e16f45a-71c1-411e-9084-c950519e768d" (UID: "7e16f45a-71c1-411e-9084-c950519e768d"). InnerVolumeSpecName "kube-api-access-whb5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.554823 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e16f45a-71c1-411e-9084-c950519e768d" (UID: "7e16f45a-71c1-411e-9084-c950519e768d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.557830 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-config-data" (OuterVolumeSpecName: "config-data") pod "7e16f45a-71c1-411e-9084-c950519e768d" (UID: "7e16f45a-71c1-411e-9084-c950519e768d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.577141 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7e16f45a-71c1-411e-9084-c950519e768d" (UID: "7e16f45a-71c1-411e-9084-c950519e768d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.599099 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whb5h\" (UniqueName: \"kubernetes.io/projected/7e16f45a-71c1-411e-9084-c950519e768d-kube-api-access-whb5h\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.599134 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.599144 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.599152 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e16f45a-71c1-411e-9084-c950519e768d-logs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:16 crc kubenswrapper[4796]: I0930 16:31:16.599161 4796 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16f45a-71c1-411e-9084-c950519e768d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.176185 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e16f45a-71c1-411e-9084-c950519e768d","Type":"ContainerDied","Data":"32d4d3ca89b37a8cabd21e2fe36a8c30348d629e3ca7a0bd0073ecd6685a5c7d"} Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.176246 4796 scope.go:117] "RemoveContainer" containerID="9e0c4fa9f51e60247a9f5b85ab6550c10ea58c4fae85253c842c7f6f8634a34f" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.177352 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.202893 4796 scope.go:117] "RemoveContainer" containerID="8923fde2eba2d93d4cf942b68a5fadeec1b42fcb138fe20b5d3238a83dffc15f" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.206482 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.217589 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.246486 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:31:17 crc kubenswrapper[4796]: E0930 16:31:17.246879 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-log" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.246895 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-log" Sep 30 16:31:17 crc kubenswrapper[4796]: E0930 16:31:17.246914 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-metadata" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.246920 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-metadata" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.247118 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-log" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.247139 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e16f45a-71c1-411e-9084-c950519e768d" containerName="nova-metadata-metadata" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.248073 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.251497 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.252143 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.262352 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.310874 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.310933 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a679bcdb-21e4-4761-adc7-1c90e79c6650-logs\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.311025 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.311091 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dh74\" (UniqueName: \"kubernetes.io/projected/a679bcdb-21e4-4761-adc7-1c90e79c6650-kube-api-access-4dh74\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.311135 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-config-data\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.413504 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dh74\" (UniqueName: \"kubernetes.io/projected/a679bcdb-21e4-4761-adc7-1c90e79c6650-kube-api-access-4dh74\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.414191 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-config-data\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.414407 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.414576 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a679bcdb-21e4-4761-adc7-1c90e79c6650-logs\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.414736 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.416575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a679bcdb-21e4-4761-adc7-1c90e79c6650-logs\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.420653 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.420957 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-config-data\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.421491 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a679bcdb-21e4-4761-adc7-1c90e79c6650-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.436122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dh74\" (UniqueName: \"kubernetes.io/projected/a679bcdb-21e4-4761-adc7-1c90e79c6650-kube-api-access-4dh74\") pod \"nova-metadata-0\" (UID: \"a679bcdb-21e4-4761-adc7-1c90e79c6650\") " pod="openstack/nova-metadata-0" Sep 30 16:31:17 crc kubenswrapper[4796]: I0930 16:31:17.565346 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 16:31:18 crc kubenswrapper[4796]: I0930 16:31:18.058706 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 16:31:18 crc kubenswrapper[4796]: W0930 16:31:18.070163 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda679bcdb_21e4_4761_adc7_1c90e79c6650.slice/crio-b8a38aef0b72cd5eb22c7b061a9dbe7c274561ef88c037090b87cad97d8a8b2d WatchSource:0}: Error finding container b8a38aef0b72cd5eb22c7b061a9dbe7c274561ef88c037090b87cad97d8a8b2d: Status 404 returned error can't find the container with id b8a38aef0b72cd5eb22c7b061a9dbe7c274561ef88c037090b87cad97d8a8b2d Sep 30 16:31:18 crc kubenswrapper[4796]: I0930 16:31:18.201537 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a679bcdb-21e4-4761-adc7-1c90e79c6650","Type":"ContainerStarted","Data":"b8a38aef0b72cd5eb22c7b061a9dbe7c274561ef88c037090b87cad97d8a8b2d"} Sep 30 16:31:18 crc kubenswrapper[4796]: I0930 16:31:18.743337 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e16f45a-71c1-411e-9084-c950519e768d" path="/var/lib/kubelet/pods/7e16f45a-71c1-411e-9084-c950519e768d/volumes" Sep 30 16:31:19 crc kubenswrapper[4796]: I0930 16:31:19.213275 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a679bcdb-21e4-4761-adc7-1c90e79c6650","Type":"ContainerStarted","Data":"70954ecd7e6297a5d62556c4affdf99419b8b8c847097009ab33231407772112"} Sep 30 16:31:19 crc kubenswrapper[4796]: I0930 16:31:19.213329 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a679bcdb-21e4-4761-adc7-1c90e79c6650","Type":"ContainerStarted","Data":"b2f57cce334b4487ab5f5c7fb27999756e30f8ac671f7931e3d82ff2a6668e21"} Sep 30 16:31:19 crc kubenswrapper[4796]: I0930 16:31:19.238130 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.238103168 podStartE2EDuration="2.238103168s" podCreationTimestamp="2025-09-30 16:31:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:31:19.232285889 +0000 UTC m=+1171.245564446" watchObservedRunningTime="2025-09-30 16:31:19.238103168 +0000 UTC m=+1171.251381735" Sep 30 16:31:19 crc kubenswrapper[4796]: I0930 16:31:19.577394 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 16:31:22 crc kubenswrapper[4796]: I0930 16:31:22.566263 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 16:31:22 crc kubenswrapper[4796]: I0930 16:31:22.566627 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 16:31:24 crc kubenswrapper[4796]: I0930 16:31:24.564512 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 16:31:24 crc kubenswrapper[4796]: I0930 16:31:24.564956 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 16:31:24 crc kubenswrapper[4796]: I0930 16:31:24.577164 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 16:31:24 crc kubenswrapper[4796]: I0930 16:31:24.627193 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 16:31:25 crc kubenswrapper[4796]: I0930 16:31:25.332958 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 16:31:25 crc kubenswrapper[4796]: I0930 16:31:25.576331 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dec95fef-d221-4739-849e-5e071ce2311a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 16:31:25 crc kubenswrapper[4796]: I0930 16:31:25.576362 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dec95fef-d221-4739-849e-5e071ce2311a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 16:31:27 crc kubenswrapper[4796]: I0930 16:31:27.566869 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 16:31:27 crc kubenswrapper[4796]: I0930 16:31:27.567221 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 16:31:28 crc kubenswrapper[4796]: I0930 16:31:28.586395 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a679bcdb-21e4-4761-adc7-1c90e79c6650" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 16:31:28 crc kubenswrapper[4796]: I0930 16:31:28.586417 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a679bcdb-21e4-4761-adc7-1c90e79c6650" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 16:31:30 crc kubenswrapper[4796]: I0930 16:31:30.287962 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 16:31:34 crc kubenswrapper[4796]: I0930 16:31:34.575161 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 16:31:34 crc kubenswrapper[4796]: I0930 16:31:34.575969 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 16:31:34 crc kubenswrapper[4796]: I0930 16:31:34.576416 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 16:31:34 crc kubenswrapper[4796]: I0930 16:31:34.576455 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 16:31:34 crc kubenswrapper[4796]: I0930 16:31:34.585873 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 16:31:34 crc kubenswrapper[4796]: I0930 16:31:34.589900 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 16:31:37 crc kubenswrapper[4796]: I0930 16:31:37.571921 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 16:31:37 crc kubenswrapper[4796]: I0930 16:31:37.576532 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 16:31:37 crc kubenswrapper[4796]: I0930 16:31:37.578798 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 16:31:38 crc kubenswrapper[4796]: I0930 16:31:38.438273 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 16:31:49 crc kubenswrapper[4796]: I0930 16:31:49.945327 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:49 crc kubenswrapper[4796]: I0930 16:31:49.947049 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-central-agent" containerID="cri-o://f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f" gracePeriod=30 Sep 30 16:31:49 crc kubenswrapper[4796]: I0930 16:31:49.947164 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="proxy-httpd" containerID="cri-o://17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c" gracePeriod=30 Sep 30 16:31:49 crc kubenswrapper[4796]: I0930 16:31:49.947176 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="sg-core" containerID="cri-o://1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4" gracePeriod=30 Sep 30 16:31:49 crc kubenswrapper[4796]: I0930 16:31:49.947342 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-notification-agent" containerID="cri-o://84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953" gracePeriod=30 Sep 30 16:31:50 crc kubenswrapper[4796]: E0930 16:31:50.125236 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd67076ba_6fad_4096_8aed_c76ad237a19a.slice/crio-conmon-1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd67076ba_6fad_4096_8aed_c76ad237a19a.slice/crio-1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd67076ba_6fad_4096_8aed_c76ad237a19a.slice/crio-conmon-17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:31:50 crc kubenswrapper[4796]: I0930 16:31:50.574642 4796 generic.go:334] "Generic (PLEG): container finished" podID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerID="17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c" exitCode=0 Sep 30 16:31:50 crc kubenswrapper[4796]: I0930 16:31:50.574672 4796 generic.go:334] "Generic (PLEG): container finished" podID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerID="1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4" exitCode=2 Sep 30 16:31:50 crc kubenswrapper[4796]: I0930 16:31:50.574679 4796 generic.go:334] "Generic (PLEG): container finished" podID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerID="f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f" exitCode=0 Sep 30 16:31:50 crc kubenswrapper[4796]: I0930 16:31:50.574682 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerDied","Data":"17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c"} Sep 30 16:31:50 crc kubenswrapper[4796]: I0930 16:31:50.574733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerDied","Data":"1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4"} Sep 30 16:31:50 crc kubenswrapper[4796]: I0930 16:31:50.574747 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerDied","Data":"f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f"} Sep 30 16:31:51 crc kubenswrapper[4796]: I0930 16:31:51.128367 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:31:51 crc kubenswrapper[4796]: I0930 16:31:51.757562 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:31:54 crc kubenswrapper[4796]: I0930 16:31:54.950183 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="rabbitmq" containerID="cri-o://7fe663e8c846cefe85de508c57afcab325db9555b0e33b961273e60d34f8f45e" gracePeriod=604797 Sep 30 16:31:55 crc kubenswrapper[4796]: I0930 16:31:55.728235 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="rabbitmq" containerID="cri-o://3f3ecd9dc8a6d7039567911cdd7c6d0fc7d9e82b6a0ea1a6e6a50ed719375835" gracePeriod=604797 Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.577574 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.671143 4796 generic.go:334] "Generic (PLEG): container finished" podID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerID="84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953" exitCode=0 Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.671197 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerDied","Data":"84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953"} Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.671229 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67076ba-6fad-4096-8aed-c76ad237a19a","Type":"ContainerDied","Data":"870540ea0b26857d8bf980a47dee42871a19bb19d798f9551f731e5919be3af3"} Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.671251 4796 scope.go:117] "RemoveContainer" containerID="17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.671327 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.697008 4796 scope.go:117] "RemoveContainer" containerID="1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706134 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-scripts\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706182 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-ceilometer-tls-certs\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706217 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-sg-core-conf-yaml\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706297 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6ktr\" (UniqueName: \"kubernetes.io/projected/d67076ba-6fad-4096-8aed-c76ad237a19a-kube-api-access-g6ktr\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706343 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-log-httpd\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706457 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-config-data\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706493 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-combined-ca-bundle\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.706545 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-run-httpd\") pod \"d67076ba-6fad-4096-8aed-c76ad237a19a\" (UID: \"d67076ba-6fad-4096-8aed-c76ad237a19a\") " Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.707117 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.707155 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.711212 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-scripts" (OuterVolumeSpecName: "scripts") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.711242 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67076ba-6fad-4096-8aed-c76ad237a19a-kube-api-access-g6ktr" (OuterVolumeSpecName: "kube-api-access-g6ktr") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "kube-api-access-g6ktr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.715666 4796 scope.go:117] "RemoveContainer" containerID="84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.737743 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.761407 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.803212 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809223 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809259 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809271 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809286 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809298 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6ktr\" (UniqueName: \"kubernetes.io/projected/d67076ba-6fad-4096-8aed-c76ad237a19a-kube-api-access-g6ktr\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809309 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67076ba-6fad-4096-8aed-c76ad237a19a-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.809321 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.821996 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-config-data" (OuterVolumeSpecName: "config-data") pod "d67076ba-6fad-4096-8aed-c76ad237a19a" (UID: "d67076ba-6fad-4096-8aed-c76ad237a19a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.828922 4796 scope.go:117] "RemoveContainer" containerID="f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.846434 4796 scope.go:117] "RemoveContainer" containerID="17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c" Sep 30 16:31:56 crc kubenswrapper[4796]: E0930 16:31:56.846875 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c\": container with ID starting with 17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c not found: ID does not exist" containerID="17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.846906 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c"} err="failed to get container status \"17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c\": rpc error: code = NotFound desc = could not find container \"17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c\": container with ID starting with 17ccce3833b9292ebe0279629049c6956043be48fab5e9ea0a3c7c9656c9b12c not found: ID does not exist" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.846928 4796 scope.go:117] "RemoveContainer" containerID="1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4" Sep 30 16:31:56 crc kubenswrapper[4796]: E0930 16:31:56.847481 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4\": container with ID starting with 1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4 not found: ID does not exist" containerID="1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.847505 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4"} err="failed to get container status \"1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4\": rpc error: code = NotFound desc = could not find container \"1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4\": container with ID starting with 1fe7f8d0887ecd853ecb08275a7a219baaccb95f42d06bd58e5f286e6632fae4 not found: ID does not exist" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.847518 4796 scope.go:117] "RemoveContainer" containerID="84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953" Sep 30 16:31:56 crc kubenswrapper[4796]: E0930 16:31:56.847734 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953\": container with ID starting with 84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953 not found: ID does not exist" containerID="84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.847768 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953"} err="failed to get container status \"84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953\": rpc error: code = NotFound desc = could not find container \"84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953\": container with ID starting with 84bdf9e68113753727a85324cfc4e5ead322d89317040f802362f23cdaef2953 not found: ID does not exist" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.847786 4796 scope.go:117] "RemoveContainer" containerID="f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f" Sep 30 16:31:56 crc kubenswrapper[4796]: E0930 16:31:56.848094 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f\": container with ID starting with f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f not found: ID does not exist" containerID="f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.848114 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f"} err="failed to get container status \"f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f\": rpc error: code = NotFound desc = could not find container \"f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f\": container with ID starting with f4f61717a6489b1cd8e219d48f9d7596dd2840f54ec1045e9c328113a522c07f not found: ID does not exist" Sep 30 16:31:56 crc kubenswrapper[4796]: I0930 16:31:56.910690 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67076ba-6fad-4096-8aed-c76ad237a19a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.018843 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.028846 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.051375 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:57 crc kubenswrapper[4796]: E0930 16:31:57.051797 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-central-agent" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.051823 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-central-agent" Sep 30 16:31:57 crc kubenswrapper[4796]: E0930 16:31:57.051866 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-notification-agent" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.051878 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-notification-agent" Sep 30 16:31:57 crc kubenswrapper[4796]: E0930 16:31:57.051889 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="sg-core" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.051900 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="sg-core" Sep 30 16:31:57 crc kubenswrapper[4796]: E0930 16:31:57.051919 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="proxy-httpd" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.051926 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="proxy-httpd" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.052179 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="sg-core" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.052208 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-notification-agent" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.052218 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="ceilometer-central-agent" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.052245 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" containerName="proxy-httpd" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.054522 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.059178 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.061600 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.061761 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.064696 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.114628 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-config-data\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.114884 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-run-httpd\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.115022 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-scripts\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.115110 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-log-httpd\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.115420 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.115462 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.115704 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkgff\" (UniqueName: \"kubernetes.io/projected/cb9a76fb-3656-4a79-96a8-650c99bac0d0-kube-api-access-jkgff\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.115761 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.216902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.216963 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.217746 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkgff\" (UniqueName: \"kubernetes.io/projected/cb9a76fb-3656-4a79-96a8-650c99bac0d0-kube-api-access-jkgff\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.217784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.217875 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-config-data\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.217934 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-run-httpd\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.217998 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-scripts\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.218021 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-log-httpd\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.218402 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-run-httpd\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.218431 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-log-httpd\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.220398 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.221227 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-config-data\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.221784 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-scripts\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.222144 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.224356 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.235612 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkgff\" (UniqueName: \"kubernetes.io/projected/cb9a76fb-3656-4a79-96a8-650c99bac0d0-kube-api-access-jkgff\") pod \"ceilometer-0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.378834 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:31:57 crc kubenswrapper[4796]: I0930 16:31:57.919780 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:31:57 crc kubenswrapper[4796]: W0930 16:31:57.925231 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb9a76fb_3656_4a79_96a8_650c99bac0d0.slice/crio-9524ad6481c288d48a368551c647202cecaaaa62676fa3f828e4904f514b68ec WatchSource:0}: Error finding container 9524ad6481c288d48a368551c647202cecaaaa62676fa3f828e4904f514b68ec: Status 404 returned error can't find the container with id 9524ad6481c288d48a368551c647202cecaaaa62676fa3f828e4904f514b68ec Sep 30 16:31:58 crc kubenswrapper[4796]: I0930 16:31:58.694587 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerStarted","Data":"9524ad6481c288d48a368551c647202cecaaaa62676fa3f828e4904f514b68ec"} Sep 30 16:31:58 crc kubenswrapper[4796]: I0930 16:31:58.744970 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d67076ba-6fad-4096-8aed-c76ad237a19a" path="/var/lib/kubelet/pods/d67076ba-6fad-4096-8aed-c76ad237a19a/volumes" Sep 30 16:32:01 crc kubenswrapper[4796]: I0930 16:32:01.738779 4796 generic.go:334] "Generic (PLEG): container finished" podID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerID="7fe663e8c846cefe85de508c57afcab325db9555b0e33b961273e60d34f8f45e" exitCode=0 Sep 30 16:32:01 crc kubenswrapper[4796]: I0930 16:32:01.738872 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73","Type":"ContainerDied","Data":"7fe663e8c846cefe85de508c57afcab325db9555b0e33b961273e60d34f8f45e"} Sep 30 16:32:02 crc kubenswrapper[4796]: I0930 16:32:02.754176 4796 generic.go:334] "Generic (PLEG): container finished" podID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerID="3f3ecd9dc8a6d7039567911cdd7c6d0fc7d9e82b6a0ea1a6e6a50ed719375835" exitCode=0 Sep 30 16:32:02 crc kubenswrapper[4796]: I0930 16:32:02.754458 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee37a44-c6b6-4d75-ad41-9ff1402925f1","Type":"ContainerDied","Data":"3f3ecd9dc8a6d7039567911cdd7c6d0fc7d9e82b6a0ea1a6e6a50ed719375835"} Sep 30 16:32:03 crc kubenswrapper[4796]: I0930 16:32:03.369817 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Sep 30 16:32:03 crc kubenswrapper[4796]: I0930 16:32:03.647401 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.132246 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-4tqgn"] Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.134051 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.136336 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.154182 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-4tqgn"] Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.274574 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.274801 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-config\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.274854 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.274885 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-svc\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.274906 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hn5p\" (UniqueName: \"kubernetes.io/projected/5511a259-1fa7-4755-830f-ee4ca2e93adb-kube-api-access-7hn5p\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.274971 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.275081 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376539 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-config\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376633 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-svc\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376657 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hn5p\" (UniqueName: \"kubernetes.io/projected/5511a259-1fa7-4755-830f-ee4ca2e93adb-kube-api-access-7hn5p\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376706 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376773 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.376810 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.377464 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.378011 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-svc\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.379878 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.380957 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.381203 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.382133 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-config\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.399708 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hn5p\" (UniqueName: \"kubernetes.io/projected/5511a259-1fa7-4755-830f-ee4ca2e93adb-kube-api-access-7hn5p\") pod \"dnsmasq-dns-67b789f86c-4tqgn\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:04 crc kubenswrapper[4796]: I0930 16:32:04.470290 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.230538 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.331081 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-erlang-cookie-secret\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.331223 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-confd\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.331247 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnj2s\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-kube-api-access-mnj2s\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.331292 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-config-data\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332046 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-pod-info\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332125 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-tls\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332170 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-erlang-cookie\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332249 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-plugins\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332279 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-server-conf\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332321 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332365 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-plugins-conf\") pod \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\" (UID: \"0ee37a44-c6b6-4d75-ad41-9ff1402925f1\") " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.332692 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.333294 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.333872 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.334470 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.337219 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.338154 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-pod-info" (OuterVolumeSpecName: "pod-info") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.339168 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-kube-api-access-mnj2s" (OuterVolumeSpecName: "kube-api-access-mnj2s") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "kube-api-access-mnj2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.340617 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.344929 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.414104 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-config-data" (OuterVolumeSpecName: "config-data") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435144 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435170 4796 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435180 4796 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435189 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnj2s\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-kube-api-access-mnj2s\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435197 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435205 4796 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435213 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.435223 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.439647 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-server-conf" (OuterVolumeSpecName: "server-conf") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.496068 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.504173 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0ee37a44-c6b6-4d75-ad41-9ff1402925f1" (UID: "0ee37a44-c6b6-4d75-ad41-9ff1402925f1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.536573 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.536612 4796 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee37a44-c6b6-4d75-ad41-9ff1402925f1-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.536623 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.815060 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee37a44-c6b6-4d75-ad41-9ff1402925f1","Type":"ContainerDied","Data":"2c7e94077a9e870b219f9a15767d5dfed0fc2a3c9697a87652be5d8c549b6179"} Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.815115 4796 scope.go:117] "RemoveContainer" containerID="3f3ecd9dc8a6d7039567911cdd7c6d0fc7d9e82b6a0ea1a6e6a50ed719375835" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.815277 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.850907 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.858841 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.880947 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:32:07 crc kubenswrapper[4796]: E0930 16:32:07.882674 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="rabbitmq" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.882703 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="rabbitmq" Sep 30 16:32:07 crc kubenswrapper[4796]: E0930 16:32:07.882715 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="setup-container" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.882724 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="setup-container" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.884398 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" containerName="rabbitmq" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.889529 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.894646 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.897144 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.904843 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gwtqv" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.904913 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.904944 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.905113 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.905362 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.909738 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943080 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943200 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943242 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943259 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943272 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943299 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943347 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943594 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943615 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:07 crc kubenswrapper[4796]: I0930 16:32:07.943705 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x6n7\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-kube-api-access-6x6n7\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.031501 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045297 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045394 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045426 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045448 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045482 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045509 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045553 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045575 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045600 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045639 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x6n7\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-kube-api-access-6x6n7\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.045781 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.046338 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.046762 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.047249 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.047535 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.048026 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.048648 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.049572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.049836 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.051796 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.051934 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.072058 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x6n7\" (UniqueName: \"kubernetes.io/projected/5d4dc01f-d8ed-4040-b4f7-53498e2cc237-kube-api-access-6x6n7\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.079383 4796 scope.go:117] "RemoveContainer" containerID="677365ad12205b095e7ab1f45f8893d7344d64c584f08a009446e2d97c40ed89" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.104483 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d4dc01f-d8ed-4040-b4f7-53498e2cc237\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147579 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-plugins\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147662 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-pod-info\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147679 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147710 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fw79\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-kube-api-access-7fw79\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147802 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-erlang-cookie\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147852 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-plugins-conf\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147870 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-config-data\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147930 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-erlang-cookie-secret\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.147953 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-server-conf\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.148038 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-tls\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.148059 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-confd\") pod \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\" (UID: \"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73\") " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.149112 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.153583 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.157395 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-pod-info" (OuterVolumeSpecName: "pod-info") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.163702 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.185173 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.198706 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.198813 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.218273 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-kube-api-access-7fw79" (OuterVolumeSpecName: "kube-api-access-7fw79") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "kube-api-access-7fw79". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.219281 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.242493 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-config-data" (OuterVolumeSpecName: "config-data") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250600 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250630 4796 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250639 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250650 4796 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250657 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250666 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250674 4796 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250699 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.250707 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fw79\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-kube-api-access-7fw79\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.292772 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.355547 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.371261 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-server-conf" (OuterVolumeSpecName: "server-conf") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.391211 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" (UID: "903f4e99-1ea4-46b3-8b5e-2ba439c5ba73"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.457951 4796 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.458007 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.758836 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ee37a44-c6b6-4d75-ad41-9ff1402925f1" path="/var/lib/kubelet/pods/0ee37a44-c6b6-4d75-ad41-9ff1402925f1/volumes" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.767052 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-4tqgn"] Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.840616 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.840624 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"903f4e99-1ea4-46b3-8b5e-2ba439c5ba73","Type":"ContainerDied","Data":"80aa571d96bf60b76d3701aa197d8c928f71f3e64058198ffd4f98c24b84ff2e"} Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.840922 4796 scope.go:117] "RemoveContainer" containerID="7fe663e8c846cefe85de508c57afcab325db9555b0e33b961273e60d34f8f45e" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.844735 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" event={"ID":"5511a259-1fa7-4755-830f-ee4ca2e93adb","Type":"ContainerStarted","Data":"98311b680c8e50e92aedba621f20ee3a373a82efadf5efbe44c11263b7ef1fb4"} Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.887505 4796 scope.go:117] "RemoveContainer" containerID="8a60af1eaba786e386562dfc364d8f6a95c91b30228085291d68c87f3723546f" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.888995 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.908613 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.920693 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.930594 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:32:08 crc kubenswrapper[4796]: E0930 16:32:08.931160 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="setup-container" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.931184 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="setup-container" Sep 30 16:32:08 crc kubenswrapper[4796]: E0930 16:32:08.931223 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="rabbitmq" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.931232 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="rabbitmq" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.931476 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" containerName="rabbitmq" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.932498 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.937538 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.937744 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-khhtv" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.937884 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.938823 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.939785 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.939854 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.940211 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.960625 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971538 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-config-data\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971616 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971639 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971671 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75cf\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-kube-api-access-s75cf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971715 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971889 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971919 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971949 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:08 crc kubenswrapper[4796]: I0930 16:32:08.971970 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073736 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073789 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073841 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073882 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073909 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-config-data\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073930 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073951 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.073994 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75cf\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-kube-api-access-s75cf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.074035 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.074053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.074179 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.077202 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.078613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.078883 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.079489 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.080296 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-config-data\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.080898 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.081125 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.081364 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.081929 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.093722 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75cf\" (UniqueName: \"kubernetes.io/projected/7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5-kube-api-access-s75cf\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.116097 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5\") " pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.269879 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.737759 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.859631 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d4dc01f-d8ed-4040-b4f7-53498e2cc237","Type":"ContainerStarted","Data":"eaece965a4ae7c23bd3e7b5ac24c8977aff3ca0267fe05af96bec61ddce20c95"} Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.859758 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d4dc01f-d8ed-4040-b4f7-53498e2cc237","Type":"ContainerStarted","Data":"eed524c35e0cea61543c1d42e17b486af5f65a007a6a284c4760ce5e66ab2c79"} Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.867381 4796 generic.go:334] "Generic (PLEG): container finished" podID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerID="f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22" exitCode=0 Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.867453 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" event={"ID":"5511a259-1fa7-4755-830f-ee4ca2e93adb","Type":"ContainerDied","Data":"f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22"} Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.883137 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerStarted","Data":"4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f"} Sep 30 16:32:09 crc kubenswrapper[4796]: I0930 16:32:09.887658 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5","Type":"ContainerStarted","Data":"b9657a7fc84e5c5d1ddab1922cc4b2dec32b380c8334d63668f8887039161f7f"} Sep 30 16:32:10 crc kubenswrapper[4796]: I0930 16:32:10.746755 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="903f4e99-1ea4-46b3-8b5e-2ba439c5ba73" path="/var/lib/kubelet/pods/903f4e99-1ea4-46b3-8b5e-2ba439c5ba73/volumes" Sep 30 16:32:10 crc kubenswrapper[4796]: I0930 16:32:10.899248 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" event={"ID":"5511a259-1fa7-4755-830f-ee4ca2e93adb","Type":"ContainerStarted","Data":"1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f"} Sep 30 16:32:10 crc kubenswrapper[4796]: I0930 16:32:10.899488 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:10 crc kubenswrapper[4796]: I0930 16:32:10.904731 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerStarted","Data":"d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57"} Sep 30 16:32:10 crc kubenswrapper[4796]: I0930 16:32:10.911295 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5","Type":"ContainerStarted","Data":"acb3590cbb0d2167a217dc230a4f421f52d6cf29c79a704609f87fa616a833df"} Sep 30 16:32:10 crc kubenswrapper[4796]: I0930 16:32:10.933694 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" podStartSLOduration=6.933670855 podStartE2EDuration="6.933670855s" podCreationTimestamp="2025-09-30 16:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:32:10.926433244 +0000 UTC m=+1222.939711771" watchObservedRunningTime="2025-09-30 16:32:10.933670855 +0000 UTC m=+1222.946949382" Sep 30 16:32:11 crc kubenswrapper[4796]: I0930 16:32:11.920050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerStarted","Data":"6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622"} Sep 30 16:32:12 crc kubenswrapper[4796]: I0930 16:32:12.937879 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerStarted","Data":"067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3"} Sep 30 16:32:12 crc kubenswrapper[4796]: I0930 16:32:12.938432 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:32:12 crc kubenswrapper[4796]: I0930 16:32:12.988625 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.518255305 podStartE2EDuration="15.988597553s" podCreationTimestamp="2025-09-30 16:31:57 +0000 UTC" firstStartedPulling="2025-09-30 16:31:57.928160172 +0000 UTC m=+1209.941438729" lastFinishedPulling="2025-09-30 16:32:12.39850244 +0000 UTC m=+1224.411780977" observedRunningTime="2025-09-30 16:32:12.966627073 +0000 UTC m=+1224.979905610" watchObservedRunningTime="2025-09-30 16:32:12.988597553 +0000 UTC m=+1225.001876120" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.472192 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.538950 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-rgqtc"] Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.539339 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerName="dnsmasq-dns" containerID="cri-o://ad143d078d8ed653b598e264505cbf207b688cc9c70df7af71005abef2d4d7ca" gracePeriod=10 Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.707258 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-f4rxb"] Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.712549 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.721212 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-f4rxb"] Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811621 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811669 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811705 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811737 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811785 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlmxq\" (UniqueName: \"kubernetes.io/projected/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-kube-api-access-dlmxq\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811880 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.811926 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-config\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914090 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914137 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914157 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914175 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlmxq\" (UniqueName: \"kubernetes.io/projected/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-kube-api-access-dlmxq\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914281 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.914326 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-config\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.915651 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.916783 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.916922 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-config\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.917389 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.917419 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.918732 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:19 crc kubenswrapper[4796]: I0930 16:32:19.936215 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlmxq\" (UniqueName: \"kubernetes.io/projected/eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d-kube-api-access-dlmxq\") pod \"dnsmasq-dns-cb6ffcf87-f4rxb\" (UID: \"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d\") " pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.030280 4796 generic.go:334] "Generic (PLEG): container finished" podID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerID="ad143d078d8ed653b598e264505cbf207b688cc9c70df7af71005abef2d4d7ca" exitCode=0 Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.030320 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" event={"ID":"02ef0f96-7297-4ba8-a571-0289dc0b4e07","Type":"ContainerDied","Data":"ad143d078d8ed653b598e264505cbf207b688cc9c70df7af71005abef2d4d7ca"} Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.030345 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" event={"ID":"02ef0f96-7297-4ba8-a571-0289dc0b4e07","Type":"ContainerDied","Data":"90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d"} Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.030356 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.045514 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.049748 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.117684 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-svc\") pod \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.118015 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n2p6\" (UniqueName: \"kubernetes.io/projected/02ef0f96-7297-4ba8-a571-0289dc0b4e07-kube-api-access-9n2p6\") pod \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.118047 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-swift-storage-0\") pod \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.118143 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-config\") pod \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.118375 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-sb\") pod \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.118431 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-nb\") pod \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\" (UID: \"02ef0f96-7297-4ba8-a571-0289dc0b4e07\") " Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.129145 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ef0f96-7297-4ba8-a571-0289dc0b4e07-kube-api-access-9n2p6" (OuterVolumeSpecName: "kube-api-access-9n2p6") pod "02ef0f96-7297-4ba8-a571-0289dc0b4e07" (UID: "02ef0f96-7297-4ba8-a571-0289dc0b4e07"). InnerVolumeSpecName "kube-api-access-9n2p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.168831 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02ef0f96-7297-4ba8-a571-0289dc0b4e07" (UID: "02ef0f96-7297-4ba8-a571-0289dc0b4e07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.168980 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "02ef0f96-7297-4ba8-a571-0289dc0b4e07" (UID: "02ef0f96-7297-4ba8-a571-0289dc0b4e07"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.178694 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-config" (OuterVolumeSpecName: "config") pod "02ef0f96-7297-4ba8-a571-0289dc0b4e07" (UID: "02ef0f96-7297-4ba8-a571-0289dc0b4e07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.196400 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "02ef0f96-7297-4ba8-a571-0289dc0b4e07" (UID: "02ef0f96-7297-4ba8-a571-0289dc0b4e07"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.216024 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "02ef0f96-7297-4ba8-a571-0289dc0b4e07" (UID: "02ef0f96-7297-4ba8-a571-0289dc0b4e07"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.220264 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.220292 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.220304 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.220312 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.220322 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n2p6\" (UniqueName: \"kubernetes.io/projected/02ef0f96-7297-4ba8-a571-0289dc0b4e07-kube-api-access-9n2p6\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.220331 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef0f96-7297-4ba8-a571-0289dc0b4e07-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:20 crc kubenswrapper[4796]: I0930 16:32:20.562399 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-f4rxb"] Sep 30 16:32:20 crc kubenswrapper[4796]: W0930 16:32:20.565078 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb4340a6_f8f3_4d95_922c_fddf4e6c5a9d.slice/crio-5b7bb941e5fa0f9e3647a0401618882abe7c4b228684841b738c1dadb3050790 WatchSource:0}: Error finding container 5b7bb941e5fa0f9e3647a0401618882abe7c4b228684841b738c1dadb3050790: Status 404 returned error can't find the container with id 5b7bb941e5fa0f9e3647a0401618882abe7c4b228684841b738c1dadb3050790 Sep 30 16:32:20 crc kubenswrapper[4796]: E0930 16:32:20.979102 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02ef0f96_7297_4ba8_a571_0289dc0b4e07.slice/crio-90b8f39b768f3fbb2872aa856a9d809ef2c3b306418ac2b7edbc2b40cd199a1d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb4340a6_f8f3_4d95_922c_fddf4e6c5a9d.slice/crio-conmon-d7273150b89b5934b0c54485aa670af82febc6d3ddeffea6d688e8c76c3a9a02.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02ef0f96_7297_4ba8_a571_0289dc0b4e07.slice\": RecentStats: unable to find data in memory cache]" Sep 30 16:32:21 crc kubenswrapper[4796]: I0930 16:32:21.047495 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d" containerID="d7273150b89b5934b0c54485aa670af82febc6d3ddeffea6d688e8c76c3a9a02" exitCode=0 Sep 30 16:32:21 crc kubenswrapper[4796]: I0930 16:32:21.047718 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" event={"ID":"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d","Type":"ContainerDied","Data":"d7273150b89b5934b0c54485aa670af82febc6d3ddeffea6d688e8c76c3a9a02"} Sep 30 16:32:21 crc kubenswrapper[4796]: I0930 16:32:21.047767 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" event={"ID":"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d","Type":"ContainerStarted","Data":"5b7bb941e5fa0f9e3647a0401618882abe7c4b228684841b738c1dadb3050790"} Sep 30 16:32:21 crc kubenswrapper[4796]: I0930 16:32:21.047962 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-rgqtc" Sep 30 16:32:21 crc kubenswrapper[4796]: I0930 16:32:21.102431 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-rgqtc"] Sep 30 16:32:21 crc kubenswrapper[4796]: I0930 16:32:21.111484 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-rgqtc"] Sep 30 16:32:22 crc kubenswrapper[4796]: I0930 16:32:22.063827 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" event={"ID":"eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d","Type":"ContainerStarted","Data":"340e6217b75cd959daa59284cece8db72371901486c05226f7eb7d3bab16a7c1"} Sep 30 16:32:22 crc kubenswrapper[4796]: I0930 16:32:22.066897 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:22 crc kubenswrapper[4796]: I0930 16:32:22.097373 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" podStartSLOduration=3.097352645 podStartE2EDuration="3.097352645s" podCreationTimestamp="2025-09-30 16:32:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:32:22.097092697 +0000 UTC m=+1234.110371274" watchObservedRunningTime="2025-09-30 16:32:22.097352645 +0000 UTC m=+1234.110631182" Sep 30 16:32:22 crc kubenswrapper[4796]: I0930 16:32:22.755467 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" path="/var/lib/kubelet/pods/02ef0f96-7297-4ba8-a571-0289dc0b4e07/volumes" Sep 30 16:32:27 crc kubenswrapper[4796]: I0930 16:32:27.403744 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.047160 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-f4rxb" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.126960 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-4tqgn"] Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.127204 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerName="dnsmasq-dns" containerID="cri-o://1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f" gracePeriod=10 Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.569764 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739144 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-openstack-edpm-ipam\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739258 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-svc\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739361 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hn5p\" (UniqueName: \"kubernetes.io/projected/5511a259-1fa7-4755-830f-ee4ca2e93adb-kube-api-access-7hn5p\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739405 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-swift-storage-0\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739430 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-config\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739469 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-nb\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.739511 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-sb\") pod \"5511a259-1fa7-4755-830f-ee4ca2e93adb\" (UID: \"5511a259-1fa7-4755-830f-ee4ca2e93adb\") " Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.746103 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5511a259-1fa7-4755-830f-ee4ca2e93adb-kube-api-access-7hn5p" (OuterVolumeSpecName: "kube-api-access-7hn5p") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "kube-api-access-7hn5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.798268 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-config" (OuterVolumeSpecName: "config") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.812378 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.819634 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.829052 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.832850 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.842615 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.842652 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hn5p\" (UniqueName: \"kubernetes.io/projected/5511a259-1fa7-4755-830f-ee4ca2e93adb-kube-api-access-7hn5p\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.842662 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.842671 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.842682 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.842692 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.846568 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5511a259-1fa7-4755-830f-ee4ca2e93adb" (UID: "5511a259-1fa7-4755-830f-ee4ca2e93adb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:32:30 crc kubenswrapper[4796]: I0930 16:32:30.944453 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5511a259-1fa7-4755-830f-ee4ca2e93adb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.096610 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.097625 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.166030 4796 generic.go:334] "Generic (PLEG): container finished" podID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerID="1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f" exitCode=0 Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.166072 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" event={"ID":"5511a259-1fa7-4755-830f-ee4ca2e93adb","Type":"ContainerDied","Data":"1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f"} Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.166099 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" event={"ID":"5511a259-1fa7-4755-830f-ee4ca2e93adb","Type":"ContainerDied","Data":"98311b680c8e50e92aedba621f20ee3a373a82efadf5efbe44c11263b7ef1fb4"} Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.166119 4796 scope.go:117] "RemoveContainer" containerID="1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.166153 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-4tqgn" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.196289 4796 scope.go:117] "RemoveContainer" containerID="f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.214875 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-4tqgn"] Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.225806 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-4tqgn"] Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.254880 4796 scope.go:117] "RemoveContainer" containerID="1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f" Sep 30 16:32:31 crc kubenswrapper[4796]: E0930 16:32:31.257824 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f\": container with ID starting with 1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f not found: ID does not exist" containerID="1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.257906 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f"} err="failed to get container status \"1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f\": rpc error: code = NotFound desc = could not find container \"1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f\": container with ID starting with 1a7fb17d7d5457582b2dd4d776ae29f19bfb5bcd8de0f6d6ad30a0ed7787907f not found: ID does not exist" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.257946 4796 scope.go:117] "RemoveContainer" containerID="f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22" Sep 30 16:32:31 crc kubenswrapper[4796]: E0930 16:32:31.258510 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22\": container with ID starting with f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22 not found: ID does not exist" containerID="f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22" Sep 30 16:32:31 crc kubenswrapper[4796]: I0930 16:32:31.258559 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22"} err="failed to get container status \"f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22\": rpc error: code = NotFound desc = could not find container \"f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22\": container with ID starting with f07b7154de91e9a68316ab1abfce2166ebb9dbd01aa4052a0ba0fbb74be1ab22 not found: ID does not exist" Sep 30 16:32:32 crc kubenswrapper[4796]: I0930 16:32:32.752826 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" path="/var/lib/kubelet/pods/5511a259-1fa7-4755-830f-ee4ca2e93adb/volumes" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.717703 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8"] Sep 30 16:32:38 crc kubenswrapper[4796]: E0930 16:32:38.719873 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerName="dnsmasq-dns" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.720029 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerName="dnsmasq-dns" Sep 30 16:32:38 crc kubenswrapper[4796]: E0930 16:32:38.720136 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerName="init" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.720222 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerName="init" Sep 30 16:32:38 crc kubenswrapper[4796]: E0930 16:32:38.720330 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerName="dnsmasq-dns" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.720421 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerName="dnsmasq-dns" Sep 30 16:32:38 crc kubenswrapper[4796]: E0930 16:32:38.720531 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerName="init" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.720607 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerName="init" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.720967 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ef0f96-7297-4ba8-a571-0289dc0b4e07" containerName="dnsmasq-dns" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.721241 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5511a259-1fa7-4755-830f-ee4ca2e93adb" containerName="dnsmasq-dns" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.722054 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.724818 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.725449 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.725735 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.726172 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.775865 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8"] Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.802897 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.803031 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7gr4\" (UniqueName: \"kubernetes.io/projected/6146b235-0d04-4d50-b28b-8531fe47a5d2-kube-api-access-s7gr4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.803099 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.803128 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.905088 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.905177 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7gr4\" (UniqueName: \"kubernetes.io/projected/6146b235-0d04-4d50-b28b-8531fe47a5d2-kube-api-access-s7gr4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.905265 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.905287 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.911127 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.914563 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.914649 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:38 crc kubenswrapper[4796]: I0930 16:32:38.938859 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7gr4\" (UniqueName: \"kubernetes.io/projected/6146b235-0d04-4d50-b28b-8531fe47a5d2-kube-api-access-s7gr4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:39 crc kubenswrapper[4796]: I0930 16:32:39.067698 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:32:39 crc kubenswrapper[4796]: I0930 16:32:39.615238 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8"] Sep 30 16:32:39 crc kubenswrapper[4796]: W0930 16:32:39.623211 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6146b235_0d04_4d50_b28b_8531fe47a5d2.slice/crio-283e4600b68cad05f924a3840e2fe72d0310de7f103141320692f397dcc11f16 WatchSource:0}: Error finding container 283e4600b68cad05f924a3840e2fe72d0310de7f103141320692f397dcc11f16: Status 404 returned error can't find the container with id 283e4600b68cad05f924a3840e2fe72d0310de7f103141320692f397dcc11f16 Sep 30 16:32:40 crc kubenswrapper[4796]: I0930 16:32:40.268172 4796 generic.go:334] "Generic (PLEG): container finished" podID="5d4dc01f-d8ed-4040-b4f7-53498e2cc237" containerID="eaece965a4ae7c23bd3e7b5ac24c8977aff3ca0267fe05af96bec61ddce20c95" exitCode=0 Sep 30 16:32:40 crc kubenswrapper[4796]: I0930 16:32:40.268266 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d4dc01f-d8ed-4040-b4f7-53498e2cc237","Type":"ContainerDied","Data":"eaece965a4ae7c23bd3e7b5ac24c8977aff3ca0267fe05af96bec61ddce20c95"} Sep 30 16:32:40 crc kubenswrapper[4796]: I0930 16:32:40.269903 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" event={"ID":"6146b235-0d04-4d50-b28b-8531fe47a5d2","Type":"ContainerStarted","Data":"283e4600b68cad05f924a3840e2fe72d0310de7f103141320692f397dcc11f16"} Sep 30 16:32:41 crc kubenswrapper[4796]: I0930 16:32:41.282871 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d4dc01f-d8ed-4040-b4f7-53498e2cc237","Type":"ContainerStarted","Data":"211dc3272ad3d817f8a7cb97ff8f61d8eaa79878d2ded2687f25f8d10f810b37"} Sep 30 16:32:41 crc kubenswrapper[4796]: I0930 16:32:41.283358 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:41 crc kubenswrapper[4796]: I0930 16:32:41.284534 4796 generic.go:334] "Generic (PLEG): container finished" podID="7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5" containerID="acb3590cbb0d2167a217dc230a4f421f52d6cf29c79a704609f87fa616a833df" exitCode=0 Sep 30 16:32:41 crc kubenswrapper[4796]: I0930 16:32:41.284589 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5","Type":"ContainerDied","Data":"acb3590cbb0d2167a217dc230a4f421f52d6cf29c79a704609f87fa616a833df"} Sep 30 16:32:41 crc kubenswrapper[4796]: I0930 16:32:41.317237 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=34.317212294 podStartE2EDuration="34.317212294s" podCreationTimestamp="2025-09-30 16:32:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:32:41.306947425 +0000 UTC m=+1253.320225952" watchObservedRunningTime="2025-09-30 16:32:41.317212294 +0000 UTC m=+1253.330490821" Sep 30 16:32:42 crc kubenswrapper[4796]: I0930 16:32:42.293627 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5","Type":"ContainerStarted","Data":"ad71a42aadac736ca91565861b1b591867128e2e18f4502feea2e8b93084cf96"} Sep 30 16:32:42 crc kubenswrapper[4796]: I0930 16:32:42.294327 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 16:32:42 crc kubenswrapper[4796]: I0930 16:32:42.315467 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=34.315451932 podStartE2EDuration="34.315451932s" podCreationTimestamp="2025-09-30 16:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:32:42.314304989 +0000 UTC m=+1254.327583516" watchObservedRunningTime="2025-09-30 16:32:42.315451932 +0000 UTC m=+1254.328730459" Sep 30 16:32:49 crc kubenswrapper[4796]: I0930 16:32:49.381067 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:32:50 crc kubenswrapper[4796]: I0930 16:32:50.374715 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" event={"ID":"6146b235-0d04-4d50-b28b-8531fe47a5d2","Type":"ContainerStarted","Data":"be0f8adcd25ae4bdc88cc7d670ad665db563096707c8ca2b1287672e224f2c73"} Sep 30 16:32:50 crc kubenswrapper[4796]: I0930 16:32:50.402786 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" podStartSLOduration=2.650066588 podStartE2EDuration="12.402765831s" podCreationTimestamp="2025-09-30 16:32:38 +0000 UTC" firstStartedPulling="2025-09-30 16:32:39.62608118 +0000 UTC m=+1251.639359707" lastFinishedPulling="2025-09-30 16:32:49.378780423 +0000 UTC m=+1261.392058950" observedRunningTime="2025-09-30 16:32:50.397324903 +0000 UTC m=+1262.410603430" watchObservedRunningTime="2025-09-30 16:32:50.402765831 +0000 UTC m=+1262.416044358" Sep 30 16:32:58 crc kubenswrapper[4796]: I0930 16:32:58.224281 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 16:32:59 crc kubenswrapper[4796]: I0930 16:32:59.273196 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 16:33:01 crc kubenswrapper[4796]: I0930 16:33:01.097303 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:33:01 crc kubenswrapper[4796]: I0930 16:33:01.097657 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:33:01 crc kubenswrapper[4796]: I0930 16:33:01.499613 4796 generic.go:334] "Generic (PLEG): container finished" podID="6146b235-0d04-4d50-b28b-8531fe47a5d2" containerID="be0f8adcd25ae4bdc88cc7d670ad665db563096707c8ca2b1287672e224f2c73" exitCode=0 Sep 30 16:33:01 crc kubenswrapper[4796]: I0930 16:33:01.499678 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" event={"ID":"6146b235-0d04-4d50-b28b-8531fe47a5d2","Type":"ContainerDied","Data":"be0f8adcd25ae4bdc88cc7d670ad665db563096707c8ca2b1287672e224f2c73"} Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.003352 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.111428 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-ssh-key\") pod \"6146b235-0d04-4d50-b28b-8531fe47a5d2\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.111590 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-repo-setup-combined-ca-bundle\") pod \"6146b235-0d04-4d50-b28b-8531fe47a5d2\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.111737 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-inventory\") pod \"6146b235-0d04-4d50-b28b-8531fe47a5d2\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.111769 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7gr4\" (UniqueName: \"kubernetes.io/projected/6146b235-0d04-4d50-b28b-8531fe47a5d2-kube-api-access-s7gr4\") pod \"6146b235-0d04-4d50-b28b-8531fe47a5d2\" (UID: \"6146b235-0d04-4d50-b28b-8531fe47a5d2\") " Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.116379 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6146b235-0d04-4d50-b28b-8531fe47a5d2" (UID: "6146b235-0d04-4d50-b28b-8531fe47a5d2"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.126959 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6146b235-0d04-4d50-b28b-8531fe47a5d2-kube-api-access-s7gr4" (OuterVolumeSpecName: "kube-api-access-s7gr4") pod "6146b235-0d04-4d50-b28b-8531fe47a5d2" (UID: "6146b235-0d04-4d50-b28b-8531fe47a5d2"). InnerVolumeSpecName "kube-api-access-s7gr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.155012 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-inventory" (OuterVolumeSpecName: "inventory") pod "6146b235-0d04-4d50-b28b-8531fe47a5d2" (UID: "6146b235-0d04-4d50-b28b-8531fe47a5d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.155844 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6146b235-0d04-4d50-b28b-8531fe47a5d2" (UID: "6146b235-0d04-4d50-b28b-8531fe47a5d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.213590 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.213618 4796 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.213632 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6146b235-0d04-4d50-b28b-8531fe47a5d2-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.213643 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7gr4\" (UniqueName: \"kubernetes.io/projected/6146b235-0d04-4d50-b28b-8531fe47a5d2-kube-api-access-s7gr4\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.523081 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" event={"ID":"6146b235-0d04-4d50-b28b-8531fe47a5d2","Type":"ContainerDied","Data":"283e4600b68cad05f924a3840e2fe72d0310de7f103141320692f397dcc11f16"} Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.523122 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="283e4600b68cad05f924a3840e2fe72d0310de7f103141320692f397dcc11f16" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.523183 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.659711 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6"] Sep 30 16:33:03 crc kubenswrapper[4796]: E0930 16:33:03.660374 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6146b235-0d04-4d50-b28b-8531fe47a5d2" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.660403 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6146b235-0d04-4d50-b28b-8531fe47a5d2" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.660761 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6146b235-0d04-4d50-b28b-8531fe47a5d2" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.661750 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.664497 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.664813 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.666427 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.666488 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.680039 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6"] Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.723222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q8s5\" (UniqueName: \"kubernetes.io/projected/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-kube-api-access-8q8s5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.723368 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.723482 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.827295 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q8s5\" (UniqueName: \"kubernetes.io/projected/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-kube-api-access-8q8s5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.827624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.827963 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.832676 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.833272 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.860055 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q8s5\" (UniqueName: \"kubernetes.io/projected/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-kube-api-access-8q8s5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-47dk6\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:03 crc kubenswrapper[4796]: I0930 16:33:03.988203 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:04 crc kubenswrapper[4796]: I0930 16:33:04.603539 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6"] Sep 30 16:33:05 crc kubenswrapper[4796]: I0930 16:33:05.547929 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" event={"ID":"80a695f7-4e2a-47f4-a4f4-724330d4e2e8","Type":"ContainerStarted","Data":"befaae3596572678605ec873ca6321a7feb88408b904ddc4039acfa00896844c"} Sep 30 16:33:07 crc kubenswrapper[4796]: I0930 16:33:07.571467 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" event={"ID":"80a695f7-4e2a-47f4-a4f4-724330d4e2e8","Type":"ContainerStarted","Data":"7b5781a06b4c4b23ae0c0e480c9a612796b11102c2fe1d082231c210d5c29c9f"} Sep 30 16:33:07 crc kubenswrapper[4796]: I0930 16:33:07.603412 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" podStartSLOduration=2.8744669480000002 podStartE2EDuration="4.603385163s" podCreationTimestamp="2025-09-30 16:33:03 +0000 UTC" firstStartedPulling="2025-09-30 16:33:04.625614102 +0000 UTC m=+1276.638892629" lastFinishedPulling="2025-09-30 16:33:06.354532277 +0000 UTC m=+1278.367810844" observedRunningTime="2025-09-30 16:33:07.589508409 +0000 UTC m=+1279.602786946" watchObservedRunningTime="2025-09-30 16:33:07.603385163 +0000 UTC m=+1279.616663710" Sep 30 16:33:09 crc kubenswrapper[4796]: I0930 16:33:09.597615 4796 generic.go:334] "Generic (PLEG): container finished" podID="80a695f7-4e2a-47f4-a4f4-724330d4e2e8" containerID="7b5781a06b4c4b23ae0c0e480c9a612796b11102c2fe1d082231c210d5c29c9f" exitCode=0 Sep 30 16:33:09 crc kubenswrapper[4796]: I0930 16:33:09.597660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" event={"ID":"80a695f7-4e2a-47f4-a4f4-724330d4e2e8","Type":"ContainerDied","Data":"7b5781a06b4c4b23ae0c0e480c9a612796b11102c2fe1d082231c210d5c29c9f"} Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.071572 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.168225 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q8s5\" (UniqueName: \"kubernetes.io/projected/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-kube-api-access-8q8s5\") pod \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.168472 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-inventory\") pod \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.168632 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-ssh-key\") pod \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\" (UID: \"80a695f7-4e2a-47f4-a4f4-724330d4e2e8\") " Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.174262 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-kube-api-access-8q8s5" (OuterVolumeSpecName: "kube-api-access-8q8s5") pod "80a695f7-4e2a-47f4-a4f4-724330d4e2e8" (UID: "80a695f7-4e2a-47f4-a4f4-724330d4e2e8"). InnerVolumeSpecName "kube-api-access-8q8s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.197217 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80a695f7-4e2a-47f4-a4f4-724330d4e2e8" (UID: "80a695f7-4e2a-47f4-a4f4-724330d4e2e8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.200734 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-inventory" (OuterVolumeSpecName: "inventory") pod "80a695f7-4e2a-47f4-a4f4-724330d4e2e8" (UID: "80a695f7-4e2a-47f4-a4f4-724330d4e2e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.273198 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q8s5\" (UniqueName: \"kubernetes.io/projected/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-kube-api-access-8q8s5\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.273230 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.273239 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80a695f7-4e2a-47f4-a4f4-724330d4e2e8-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.619909 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" event={"ID":"80a695f7-4e2a-47f4-a4f4-724330d4e2e8","Type":"ContainerDied","Data":"befaae3596572678605ec873ca6321a7feb88408b904ddc4039acfa00896844c"} Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.620284 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="befaae3596572678605ec873ca6321a7feb88408b904ddc4039acfa00896844c" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.619999 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-47dk6" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.838002 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj"] Sep 30 16:33:11 crc kubenswrapper[4796]: E0930 16:33:11.838378 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a695f7-4e2a-47f4-a4f4-724330d4e2e8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.838394 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a695f7-4e2a-47f4-a4f4-724330d4e2e8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.838630 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a695f7-4e2a-47f4-a4f4-724330d4e2e8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.839283 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.843286 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.843433 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.843666 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.843871 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.854667 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj"] Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.893014 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.893157 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x886v\" (UniqueName: \"kubernetes.io/projected/c62ce10b-8219-4394-b6c4-4514b202da42-kube-api-access-x886v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.893195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.893293 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.995270 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x886v\" (UniqueName: \"kubernetes.io/projected/c62ce10b-8219-4394-b6c4-4514b202da42-kube-api-access-x886v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.995891 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.996862 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:11 crc kubenswrapper[4796]: I0930 16:33:11.997015 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:12 crc kubenswrapper[4796]: I0930 16:33:12.001267 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:12 crc kubenswrapper[4796]: I0930 16:33:12.001871 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:12 crc kubenswrapper[4796]: I0930 16:33:12.003096 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:12 crc kubenswrapper[4796]: I0930 16:33:12.011423 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x886v\" (UniqueName: \"kubernetes.io/projected/c62ce10b-8219-4394-b6c4-4514b202da42-kube-api-access-x886v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:12 crc kubenswrapper[4796]: I0930 16:33:12.159588 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:33:12 crc kubenswrapper[4796]: I0930 16:33:12.697811 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj"] Sep 30 16:33:13 crc kubenswrapper[4796]: I0930 16:33:13.647428 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" event={"ID":"c62ce10b-8219-4394-b6c4-4514b202da42","Type":"ContainerStarted","Data":"251cafbcd91162b9a88b2fa32be458740dd1c53ca24a40ddd5b09696a3788477"} Sep 30 16:33:13 crc kubenswrapper[4796]: I0930 16:33:13.647796 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" event={"ID":"c62ce10b-8219-4394-b6c4-4514b202da42","Type":"ContainerStarted","Data":"ebc41d7bb478b68565e0571f458c1dbde517c62af08751cfe363a329d51baa98"} Sep 30 16:33:13 crc kubenswrapper[4796]: I0930 16:33:13.675085 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" podStartSLOduration=2.268590331 podStartE2EDuration="2.675062816s" podCreationTimestamp="2025-09-30 16:33:11 +0000 UTC" firstStartedPulling="2025-09-30 16:33:12.694201595 +0000 UTC m=+1284.707480122" lastFinishedPulling="2025-09-30 16:33:13.10067406 +0000 UTC m=+1285.113952607" observedRunningTime="2025-09-30 16:33:13.665083535 +0000 UTC m=+1285.678371923" watchObservedRunningTime="2025-09-30 16:33:13.675062816 +0000 UTC m=+1285.688341343" Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.097435 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.098068 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.098125 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.098917 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b7ed0c28fbd614f1b48f9a4c2dd1b94dd52891d55126b94c8724c34b2b684f2"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.099027 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://7b7ed0c28fbd614f1b48f9a4c2dd1b94dd52891d55126b94c8724c34b2b684f2" gracePeriod=600 Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.837347 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="7b7ed0c28fbd614f1b48f9a4c2dd1b94dd52891d55126b94c8724c34b2b684f2" exitCode=0 Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.837413 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"7b7ed0c28fbd614f1b48f9a4c2dd1b94dd52891d55126b94c8724c34b2b684f2"} Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.838023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2"} Sep 30 16:33:31 crc kubenswrapper[4796]: I0930 16:33:31.838053 4796 scope.go:117] "RemoveContainer" containerID="f24d81f5faa8f6182a7d380f63e2c090bd17e98f5259e863fa184efdaf83296f" Sep 30 16:33:52 crc kubenswrapper[4796]: I0930 16:33:52.598118 4796 scope.go:117] "RemoveContainer" containerID="fbcbd1746e285093161a27390765f3bb40791e8f88032b87aefdb854451cfff4" Sep 30 16:33:52 crc kubenswrapper[4796]: I0930 16:33:52.639835 4796 scope.go:117] "RemoveContainer" containerID="e124763efe374d1bbde6afc28d9f5fa5afe34f659f83cad82cbf5c6326344897" Sep 30 16:33:52 crc kubenswrapper[4796]: I0930 16:33:52.702621 4796 scope.go:117] "RemoveContainer" containerID="65984ad7c085611548307dba3be32692dc3dd06bfca53aeab4de5333bc3811c2" Sep 30 16:33:52 crc kubenswrapper[4796]: I0930 16:33:52.745288 4796 scope.go:117] "RemoveContainer" containerID="34a40a88436b6e90088af035949673031b8e24b425677f98cb88ecb849690d1c" Sep 30 16:33:52 crc kubenswrapper[4796]: I0930 16:33:52.768967 4796 scope.go:117] "RemoveContainer" containerID="8730bbebc1c3d4029c8c09cd2c745527f425996209fabb1d61c59cc744923acc" Sep 30 16:33:52 crc kubenswrapper[4796]: I0930 16:33:52.796089 4796 scope.go:117] "RemoveContainer" containerID="8f5810f1a3005e3916c0b71ad9dcde96bf6391791c015e3321790becfeb2de13" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.645518 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9rc"] Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.651780 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.656045 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9rc"] Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.751436 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpwk5\" (UniqueName: \"kubernetes.io/projected/6ffeb76a-aa97-40b8-9014-a3be6f8831db-kube-api-access-lpwk5\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.751531 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-catalog-content\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.751606 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-utilities\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.853282 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-utilities\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.853437 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpwk5\" (UniqueName: \"kubernetes.io/projected/6ffeb76a-aa97-40b8-9014-a3be6f8831db-kube-api-access-lpwk5\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.853493 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-catalog-content\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.854093 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-catalog-content\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.854110 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-utilities\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.878065 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpwk5\" (UniqueName: \"kubernetes.io/projected/6ffeb76a-aa97-40b8-9014-a3be6f8831db-kube-api-access-lpwk5\") pod \"redhat-marketplace-rt9rc\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:45 crc kubenswrapper[4796]: I0930 16:34:45.985506 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:46 crc kubenswrapper[4796]: I0930 16:34:46.449710 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9rc"] Sep 30 16:34:46 crc kubenswrapper[4796]: I0930 16:34:46.707679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerStarted","Data":"abbcc15c0c718d5eb2285993774193f3aa81919e141b26a57354c3cdafa8fe1f"} Sep 30 16:34:47 crc kubenswrapper[4796]: I0930 16:34:47.722962 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerID="dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b" exitCode=0 Sep 30 16:34:47 crc kubenswrapper[4796]: I0930 16:34:47.723069 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerDied","Data":"dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b"} Sep 30 16:34:48 crc kubenswrapper[4796]: I0930 16:34:48.750792 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerStarted","Data":"ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1"} Sep 30 16:34:49 crc kubenswrapper[4796]: I0930 16:34:49.771376 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerID="ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1" exitCode=0 Sep 30 16:34:49 crc kubenswrapper[4796]: I0930 16:34:49.771512 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerDied","Data":"ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1"} Sep 30 16:34:51 crc kubenswrapper[4796]: I0930 16:34:51.800939 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerStarted","Data":"966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe"} Sep 30 16:34:51 crc kubenswrapper[4796]: I0930 16:34:51.823630 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rt9rc" podStartSLOduration=3.963069486 podStartE2EDuration="6.823603799s" podCreationTimestamp="2025-09-30 16:34:45 +0000 UTC" firstStartedPulling="2025-09-30 16:34:47.72747646 +0000 UTC m=+1379.740754997" lastFinishedPulling="2025-09-30 16:34:50.588010783 +0000 UTC m=+1382.601289310" observedRunningTime="2025-09-30 16:34:51.816319036 +0000 UTC m=+1383.829597573" watchObservedRunningTime="2025-09-30 16:34:51.823603799 +0000 UTC m=+1383.836882346" Sep 30 16:34:52 crc kubenswrapper[4796]: I0930 16:34:52.907165 4796 scope.go:117] "RemoveContainer" containerID="adf4294f88db5a64ddda486bd1e7f186a709344f4fe70a74a1862529ec398b55" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.148649 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bkdd7"] Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.162570 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.167969 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bkdd7"] Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.254087 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdzbn\" (UniqueName: \"kubernetes.io/projected/392ebd6c-1211-4357-a605-ec8c5fe76337-kube-api-access-gdzbn\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.254135 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-utilities\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.254318 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-catalog-content\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.356091 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdzbn\" (UniqueName: \"kubernetes.io/projected/392ebd6c-1211-4357-a605-ec8c5fe76337-kube-api-access-gdzbn\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.356385 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-utilities\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.356442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-catalog-content\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.356949 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-catalog-content\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.357584 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-utilities\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: E0930 16:34:55.378067 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-conmon-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.381556 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdzbn\" (UniqueName: \"kubernetes.io/projected/392ebd6c-1211-4357-a605-ec8c5fe76337-kube-api-access-gdzbn\") pod \"redhat-operators-bkdd7\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.496322 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.966252 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bkdd7"] Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.986177 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:55 crc kubenswrapper[4796]: I0930 16:34:55.986228 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:56 crc kubenswrapper[4796]: I0930 16:34:56.047922 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:56 crc kubenswrapper[4796]: I0930 16:34:56.862567 4796 generic.go:334] "Generic (PLEG): container finished" podID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerID="3f039a86e401429878c93499cf246807989661b44eded65d71b2bfc3b7e90787" exitCode=0 Sep 30 16:34:56 crc kubenswrapper[4796]: I0930 16:34:56.862632 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkdd7" event={"ID":"392ebd6c-1211-4357-a605-ec8c5fe76337","Type":"ContainerDied","Data":"3f039a86e401429878c93499cf246807989661b44eded65d71b2bfc3b7e90787"} Sep 30 16:34:56 crc kubenswrapper[4796]: I0930 16:34:56.862986 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkdd7" event={"ID":"392ebd6c-1211-4357-a605-ec8c5fe76337","Type":"ContainerStarted","Data":"ba675e3c19348113664a6a586fcd129b4192e54d410ad59a0f9fe88f41bdef07"} Sep 30 16:34:56 crc kubenswrapper[4796]: I0930 16:34:56.943124 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:58 crc kubenswrapper[4796]: I0930 16:34:58.312153 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9rc"] Sep 30 16:34:58 crc kubenswrapper[4796]: I0930 16:34:58.889060 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rt9rc" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="registry-server" containerID="cri-o://966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe" gracePeriod=2 Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.365451 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.445149 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpwk5\" (UniqueName: \"kubernetes.io/projected/6ffeb76a-aa97-40b8-9014-a3be6f8831db-kube-api-access-lpwk5\") pod \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.445221 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-utilities\") pod \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.445372 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-catalog-content\") pod \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\" (UID: \"6ffeb76a-aa97-40b8-9014-a3be6f8831db\") " Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.447048 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-utilities" (OuterVolumeSpecName: "utilities") pod "6ffeb76a-aa97-40b8-9014-a3be6f8831db" (UID: "6ffeb76a-aa97-40b8-9014-a3be6f8831db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.451633 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ffeb76a-aa97-40b8-9014-a3be6f8831db-kube-api-access-lpwk5" (OuterVolumeSpecName: "kube-api-access-lpwk5") pod "6ffeb76a-aa97-40b8-9014-a3be6f8831db" (UID: "6ffeb76a-aa97-40b8-9014-a3be6f8831db"). InnerVolumeSpecName "kube-api-access-lpwk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.464407 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ffeb76a-aa97-40b8-9014-a3be6f8831db" (UID: "6ffeb76a-aa97-40b8-9014-a3be6f8831db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.547272 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpwk5\" (UniqueName: \"kubernetes.io/projected/6ffeb76a-aa97-40b8-9014-a3be6f8831db-kube-api-access-lpwk5\") on node \"crc\" DevicePath \"\"" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.547304 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.547314 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffeb76a-aa97-40b8-9014-a3be6f8831db-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.899462 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerID="966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe" exitCode=0 Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.899543 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerDied","Data":"966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe"} Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.899553 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9rc" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.899580 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9rc" event={"ID":"6ffeb76a-aa97-40b8-9014-a3be6f8831db","Type":"ContainerDied","Data":"abbcc15c0c718d5eb2285993774193f3aa81919e141b26a57354c3cdafa8fe1f"} Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.899604 4796 scope.go:117] "RemoveContainer" containerID="966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.903137 4796 generic.go:334] "Generic (PLEG): container finished" podID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerID="71aca77e8fb729d6597f8a37c743a6b1880369f53caf140fdcba5a93bea7d9d1" exitCode=0 Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.903171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkdd7" event={"ID":"392ebd6c-1211-4357-a605-ec8c5fe76337","Type":"ContainerDied","Data":"71aca77e8fb729d6597f8a37c743a6b1880369f53caf140fdcba5a93bea7d9d1"} Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.935687 4796 scope.go:117] "RemoveContainer" containerID="ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1" Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.958650 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9rc"] Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.969830 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9rc"] Sep 30 16:34:59 crc kubenswrapper[4796]: I0930 16:34:59.980043 4796 scope.go:117] "RemoveContainer" containerID="dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.008352 4796 scope.go:117] "RemoveContainer" containerID="966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe" Sep 30 16:35:00 crc kubenswrapper[4796]: E0930 16:35:00.008720 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe\": container with ID starting with 966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe not found: ID does not exist" containerID="966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.008751 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe"} err="failed to get container status \"966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe\": rpc error: code = NotFound desc = could not find container \"966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe\": container with ID starting with 966c2d74b244009a5f6aaa71364746b444a2026a8d00473ea8a9857b8b5a3dfe not found: ID does not exist" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.008776 4796 scope.go:117] "RemoveContainer" containerID="ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1" Sep 30 16:35:00 crc kubenswrapper[4796]: E0930 16:35:00.009514 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1\": container with ID starting with ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1 not found: ID does not exist" containerID="ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.009543 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1"} err="failed to get container status \"ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1\": rpc error: code = NotFound desc = could not find container \"ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1\": container with ID starting with ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1 not found: ID does not exist" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.009561 4796 scope.go:117] "RemoveContainer" containerID="dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b" Sep 30 16:35:00 crc kubenswrapper[4796]: E0930 16:35:00.009814 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b\": container with ID starting with dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b not found: ID does not exist" containerID="dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.009837 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b"} err="failed to get container status \"dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b\": rpc error: code = NotFound desc = could not find container \"dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b\": container with ID starting with dd2ceb4c1e058d0c27d61ec36da55f769712c5677d6846389abfcce87ce9353b not found: ID does not exist" Sep 30 16:35:00 crc kubenswrapper[4796]: I0930 16:35:00.758844 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" path="/var/lib/kubelet/pods/6ffeb76a-aa97-40b8-9014-a3be6f8831db/volumes" Sep 30 16:35:01 crc kubenswrapper[4796]: I0930 16:35:01.930017 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkdd7" event={"ID":"392ebd6c-1211-4357-a605-ec8c5fe76337","Type":"ContainerStarted","Data":"7dd05f77c0a5e66b6021d4cce406dda197589a232df375e5cea5aa8f59debc68"} Sep 30 16:35:01 crc kubenswrapper[4796]: I0930 16:35:01.956567 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bkdd7" podStartSLOduration=3.085016892 podStartE2EDuration="6.956539347s" podCreationTimestamp="2025-09-30 16:34:55 +0000 UTC" firstStartedPulling="2025-09-30 16:34:56.86736713 +0000 UTC m=+1388.880645687" lastFinishedPulling="2025-09-30 16:35:00.738889595 +0000 UTC m=+1392.752168142" observedRunningTime="2025-09-30 16:35:01.945226147 +0000 UTC m=+1393.958504754" watchObservedRunningTime="2025-09-30 16:35:01.956539347 +0000 UTC m=+1393.969817894" Sep 30 16:35:05 crc kubenswrapper[4796]: I0930 16:35:05.496692 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:35:05 crc kubenswrapper[4796]: I0930 16:35:05.497561 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:35:05 crc kubenswrapper[4796]: E0930 16:35:05.644505 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-conmon-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:35:06 crc kubenswrapper[4796]: I0930 16:35:06.545824 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bkdd7" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="registry-server" probeResult="failure" output=< Sep 30 16:35:06 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Sep 30 16:35:06 crc kubenswrapper[4796]: > Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.286343 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x79gm"] Sep 30 16:35:09 crc kubenswrapper[4796]: E0930 16:35:09.287180 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="extract-utilities" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.287195 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="extract-utilities" Sep 30 16:35:09 crc kubenswrapper[4796]: E0930 16:35:09.287208 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="extract-content" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.287214 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="extract-content" Sep 30 16:35:09 crc kubenswrapper[4796]: E0930 16:35:09.287236 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="registry-server" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.287242 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="registry-server" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.287428 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ffeb76a-aa97-40b8-9014-a3be6f8831db" containerName="registry-server" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.288930 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.309564 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x79gm"] Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.442951 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-catalog-content\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.443127 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-utilities\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.443162 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdjfx\" (UniqueName: \"kubernetes.io/projected/e4ca5376-a635-4ce5-9089-e478f4534e88-kube-api-access-cdjfx\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.544483 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-catalog-content\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.544624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-utilities\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.544650 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdjfx\" (UniqueName: \"kubernetes.io/projected/e4ca5376-a635-4ce5-9089-e478f4534e88-kube-api-access-cdjfx\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.545045 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-catalog-content\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.545146 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-utilities\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.580440 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdjfx\" (UniqueName: \"kubernetes.io/projected/e4ca5376-a635-4ce5-9089-e478f4534e88-kube-api-access-cdjfx\") pod \"community-operators-x79gm\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:09 crc kubenswrapper[4796]: I0930 16:35:09.628592 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:10 crc kubenswrapper[4796]: W0930 16:35:10.171304 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4ca5376_a635_4ce5_9089_e478f4534e88.slice/crio-2874e1045bd3fb953ac5831f1bc3139d83a7c07f891a43d07b69d798049bd5e0 WatchSource:0}: Error finding container 2874e1045bd3fb953ac5831f1bc3139d83a7c07f891a43d07b69d798049bd5e0: Status 404 returned error can't find the container with id 2874e1045bd3fb953ac5831f1bc3139d83a7c07f891a43d07b69d798049bd5e0 Sep 30 16:35:10 crc kubenswrapper[4796]: I0930 16:35:10.171530 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x79gm"] Sep 30 16:35:11 crc kubenswrapper[4796]: I0930 16:35:11.026536 4796 generic.go:334] "Generic (PLEG): container finished" podID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerID="50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99" exitCode=0 Sep 30 16:35:11 crc kubenswrapper[4796]: I0930 16:35:11.026670 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerDied","Data":"50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99"} Sep 30 16:35:11 crc kubenswrapper[4796]: I0930 16:35:11.026881 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerStarted","Data":"2874e1045bd3fb953ac5831f1bc3139d83a7c07f891a43d07b69d798049bd5e0"} Sep 30 16:35:12 crc kubenswrapper[4796]: I0930 16:35:12.048104 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerStarted","Data":"e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a"} Sep 30 16:35:13 crc kubenswrapper[4796]: I0930 16:35:13.063371 4796 generic.go:334] "Generic (PLEG): container finished" podID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerID="e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a" exitCode=0 Sep 30 16:35:13 crc kubenswrapper[4796]: I0930 16:35:13.063442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerDied","Data":"e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a"} Sep 30 16:35:14 crc kubenswrapper[4796]: I0930 16:35:14.073492 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerStarted","Data":"b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27"} Sep 30 16:35:14 crc kubenswrapper[4796]: I0930 16:35:14.096540 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x79gm" podStartSLOduration=2.386383708 podStartE2EDuration="5.096519143s" podCreationTimestamp="2025-09-30 16:35:09 +0000 UTC" firstStartedPulling="2025-09-30 16:35:11.031705128 +0000 UTC m=+1403.044983665" lastFinishedPulling="2025-09-30 16:35:13.741840553 +0000 UTC m=+1405.755119100" observedRunningTime="2025-09-30 16:35:14.091772435 +0000 UTC m=+1406.105050962" watchObservedRunningTime="2025-09-30 16:35:14.096519143 +0000 UTC m=+1406.109797670" Sep 30 16:35:15 crc kubenswrapper[4796]: I0930 16:35:15.548643 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:35:15 crc kubenswrapper[4796]: I0930 16:35:15.608794 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:35:15 crc kubenswrapper[4796]: E0930 16:35:15.877077 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-conmon-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:35:17 crc kubenswrapper[4796]: I0930 16:35:17.637707 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bkdd7"] Sep 30 16:35:17 crc kubenswrapper[4796]: I0930 16:35:17.637916 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bkdd7" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="registry-server" containerID="cri-o://7dd05f77c0a5e66b6021d4cce406dda197589a232df375e5cea5aa8f59debc68" gracePeriod=2 Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.121498 4796 generic.go:334] "Generic (PLEG): container finished" podID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerID="7dd05f77c0a5e66b6021d4cce406dda197589a232df375e5cea5aa8f59debc68" exitCode=0 Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.121579 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkdd7" event={"ID":"392ebd6c-1211-4357-a605-ec8c5fe76337","Type":"ContainerDied","Data":"7dd05f77c0a5e66b6021d4cce406dda197589a232df375e5cea5aa8f59debc68"} Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.121859 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkdd7" event={"ID":"392ebd6c-1211-4357-a605-ec8c5fe76337","Type":"ContainerDied","Data":"ba675e3c19348113664a6a586fcd129b4192e54d410ad59a0f9fe88f41bdef07"} Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.121879 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba675e3c19348113664a6a586fcd129b4192e54d410ad59a0f9fe88f41bdef07" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.185186 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.327587 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-utilities\") pod \"392ebd6c-1211-4357-a605-ec8c5fe76337\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.328122 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-catalog-content\") pod \"392ebd6c-1211-4357-a605-ec8c5fe76337\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.328191 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdzbn\" (UniqueName: \"kubernetes.io/projected/392ebd6c-1211-4357-a605-ec8c5fe76337-kube-api-access-gdzbn\") pod \"392ebd6c-1211-4357-a605-ec8c5fe76337\" (UID: \"392ebd6c-1211-4357-a605-ec8c5fe76337\") " Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.329034 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-utilities" (OuterVolumeSpecName: "utilities") pod "392ebd6c-1211-4357-a605-ec8c5fe76337" (UID: "392ebd6c-1211-4357-a605-ec8c5fe76337"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.335503 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392ebd6c-1211-4357-a605-ec8c5fe76337-kube-api-access-gdzbn" (OuterVolumeSpecName: "kube-api-access-gdzbn") pod "392ebd6c-1211-4357-a605-ec8c5fe76337" (UID: "392ebd6c-1211-4357-a605-ec8c5fe76337"). InnerVolumeSpecName "kube-api-access-gdzbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.398750 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "392ebd6c-1211-4357-a605-ec8c5fe76337" (UID: "392ebd6c-1211-4357-a605-ec8c5fe76337"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.430884 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.430915 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdzbn\" (UniqueName: \"kubernetes.io/projected/392ebd6c-1211-4357-a605-ec8c5fe76337-kube-api-access-gdzbn\") on node \"crc\" DevicePath \"\"" Sep 30 16:35:18 crc kubenswrapper[4796]: I0930 16:35:18.430928 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/392ebd6c-1211-4357-a605-ec8c5fe76337-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:35:19 crc kubenswrapper[4796]: I0930 16:35:19.130384 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkdd7" Sep 30 16:35:19 crc kubenswrapper[4796]: I0930 16:35:19.169919 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bkdd7"] Sep 30 16:35:19 crc kubenswrapper[4796]: I0930 16:35:19.180176 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bkdd7"] Sep 30 16:35:19 crc kubenswrapper[4796]: I0930 16:35:19.629079 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:19 crc kubenswrapper[4796]: I0930 16:35:19.629398 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:19 crc kubenswrapper[4796]: I0930 16:35:19.681261 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:20 crc kubenswrapper[4796]: I0930 16:35:20.192722 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:20 crc kubenswrapper[4796]: I0930 16:35:20.744693 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" path="/var/lib/kubelet/pods/392ebd6c-1211-4357-a605-ec8c5fe76337/volumes" Sep 30 16:35:21 crc kubenswrapper[4796]: I0930 16:35:21.844289 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x79gm"] Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.160484 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x79gm" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="registry-server" containerID="cri-o://b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27" gracePeriod=2 Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.642710 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.712615 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-catalog-content\") pod \"e4ca5376-a635-4ce5-9089-e478f4534e88\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.712672 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdjfx\" (UniqueName: \"kubernetes.io/projected/e4ca5376-a635-4ce5-9089-e478f4534e88-kube-api-access-cdjfx\") pod \"e4ca5376-a635-4ce5-9089-e478f4534e88\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.712726 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-utilities\") pod \"e4ca5376-a635-4ce5-9089-e478f4534e88\" (UID: \"e4ca5376-a635-4ce5-9089-e478f4534e88\") " Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.714288 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-utilities" (OuterVolumeSpecName: "utilities") pod "e4ca5376-a635-4ce5-9089-e478f4534e88" (UID: "e4ca5376-a635-4ce5-9089-e478f4534e88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.722875 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ca5376-a635-4ce5-9089-e478f4534e88-kube-api-access-cdjfx" (OuterVolumeSpecName: "kube-api-access-cdjfx") pod "e4ca5376-a635-4ce5-9089-e478f4534e88" (UID: "e4ca5376-a635-4ce5-9089-e478f4534e88"). InnerVolumeSpecName "kube-api-access-cdjfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.756625 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4ca5376-a635-4ce5-9089-e478f4534e88" (UID: "e4ca5376-a635-4ce5-9089-e478f4534e88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.815128 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.815159 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdjfx\" (UniqueName: \"kubernetes.io/projected/e4ca5376-a635-4ce5-9089-e478f4534e88-kube-api-access-cdjfx\") on node \"crc\" DevicePath \"\"" Sep 30 16:35:22 crc kubenswrapper[4796]: I0930 16:35:22.815171 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ca5376-a635-4ce5-9089-e478f4534e88-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.171530 4796 generic.go:334] "Generic (PLEG): container finished" podID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerID="b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27" exitCode=0 Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.171594 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79gm" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.171646 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerDied","Data":"b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27"} Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.172018 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79gm" event={"ID":"e4ca5376-a635-4ce5-9089-e478f4534e88","Type":"ContainerDied","Data":"2874e1045bd3fb953ac5831f1bc3139d83a7c07f891a43d07b69d798049bd5e0"} Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.172054 4796 scope.go:117] "RemoveContainer" containerID="b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.219072 4796 scope.go:117] "RemoveContainer" containerID="e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.247213 4796 scope.go:117] "RemoveContainer" containerID="50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.248765 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x79gm"] Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.256773 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x79gm"] Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.311884 4796 scope.go:117] "RemoveContainer" containerID="b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27" Sep 30 16:35:23 crc kubenswrapper[4796]: E0930 16:35:23.312593 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27\": container with ID starting with b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27 not found: ID does not exist" containerID="b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.312635 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27"} err="failed to get container status \"b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27\": rpc error: code = NotFound desc = could not find container \"b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27\": container with ID starting with b1866dd6a0d3b614461c4d0a7be867e5d673e61ba38e5f746a619d38450a2a27 not found: ID does not exist" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.312661 4796 scope.go:117] "RemoveContainer" containerID="e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a" Sep 30 16:35:23 crc kubenswrapper[4796]: E0930 16:35:23.313047 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a\": container with ID starting with e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a not found: ID does not exist" containerID="e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.313132 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a"} err="failed to get container status \"e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a\": rpc error: code = NotFound desc = could not find container \"e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a\": container with ID starting with e379c2c45897586c62f56de4cbcb6f982f67f12fda6bc969473a1d13f11b8e1a not found: ID does not exist" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.313208 4796 scope.go:117] "RemoveContainer" containerID="50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99" Sep 30 16:35:23 crc kubenswrapper[4796]: E0930 16:35:23.313624 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99\": container with ID starting with 50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99 not found: ID does not exist" containerID="50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99" Sep 30 16:35:23 crc kubenswrapper[4796]: I0930 16:35:23.313702 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99"} err="failed to get container status \"50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99\": rpc error: code = NotFound desc = could not find container \"50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99\": container with ID starting with 50288ea4d7a1874b6411dca9dba39f68d32f302ddc2bebde7a8bc1c5851edc99 not found: ID does not exist" Sep 30 16:35:24 crc kubenswrapper[4796]: I0930 16:35:24.748728 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" path="/var/lib/kubelet/pods/e4ca5376-a635-4ce5-9089-e478f4534e88/volumes" Sep 30 16:35:26 crc kubenswrapper[4796]: E0930 16:35:26.149609 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-conmon-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:35:31 crc kubenswrapper[4796]: I0930 16:35:31.097151 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:35:31 crc kubenswrapper[4796]: I0930 16:35:31.097770 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:35:36 crc kubenswrapper[4796]: E0930 16:35:36.382012 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-conmon-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:35:46 crc kubenswrapper[4796]: E0930 16:35:46.625013 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-conmon-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffeb76a_aa97_40b8_9014_a3be6f8831db.slice/crio-ca4de7fc9fc7abc1fa5caaac2c148b0b2c8b52045a89e2eec8d3b231bf3366d1.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:35:52 crc kubenswrapper[4796]: I0930 16:35:52.993463 4796 scope.go:117] "RemoveContainer" containerID="9164b7ae2ac89c441efdcec484337765d4f5f7bbed529027b453b7699d2aa2c2" Sep 30 16:35:53 crc kubenswrapper[4796]: I0930 16:35:53.019037 4796 scope.go:117] "RemoveContainer" containerID="5c49a0f66e10eb19d3d33f9e3c347b6a83533fe38d06711dfd879da28e90d6ab" Sep 30 16:35:53 crc kubenswrapper[4796]: I0930 16:35:53.078881 4796 scope.go:117] "RemoveContainer" containerID="f815e939dbd769c0072cdbf2ccfe1a282f4b085553f0457800f6dbd978a01e5e" Sep 30 16:35:53 crc kubenswrapper[4796]: I0930 16:35:53.101421 4796 scope.go:117] "RemoveContainer" containerID="ff35bcca93179fbc88089a1b50bdb3d0c7e8914f248e473a35a753d3d71d62e6" Sep 30 16:36:01 crc kubenswrapper[4796]: I0930 16:36:01.096903 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:36:01 crc kubenswrapper[4796]: I0930 16:36:01.097425 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:36:23 crc kubenswrapper[4796]: I0930 16:36:23.840861 4796 generic.go:334] "Generic (PLEG): container finished" podID="c62ce10b-8219-4394-b6c4-4514b202da42" containerID="251cafbcd91162b9a88b2fa32be458740dd1c53ca24a40ddd5b09696a3788477" exitCode=0 Sep 30 16:36:23 crc kubenswrapper[4796]: I0930 16:36:23.840889 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" event={"ID":"c62ce10b-8219-4394-b6c4-4514b202da42","Type":"ContainerDied","Data":"251cafbcd91162b9a88b2fa32be458740dd1c53ca24a40ddd5b09696a3788477"} Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.235430 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.295532 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x886v\" (UniqueName: \"kubernetes.io/projected/c62ce10b-8219-4394-b6c4-4514b202da42-kube-api-access-x886v\") pod \"c62ce10b-8219-4394-b6c4-4514b202da42\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.295610 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-inventory\") pod \"c62ce10b-8219-4394-b6c4-4514b202da42\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.295834 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-bootstrap-combined-ca-bundle\") pod \"c62ce10b-8219-4394-b6c4-4514b202da42\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.295867 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-ssh-key\") pod \"c62ce10b-8219-4394-b6c4-4514b202da42\" (UID: \"c62ce10b-8219-4394-b6c4-4514b202da42\") " Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.313941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c62ce10b-8219-4394-b6c4-4514b202da42-kube-api-access-x886v" (OuterVolumeSpecName: "kube-api-access-x886v") pod "c62ce10b-8219-4394-b6c4-4514b202da42" (UID: "c62ce10b-8219-4394-b6c4-4514b202da42"). InnerVolumeSpecName "kube-api-access-x886v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.322140 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c62ce10b-8219-4394-b6c4-4514b202da42" (UID: "c62ce10b-8219-4394-b6c4-4514b202da42"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.350389 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-inventory" (OuterVolumeSpecName: "inventory") pod "c62ce10b-8219-4394-b6c4-4514b202da42" (UID: "c62ce10b-8219-4394-b6c4-4514b202da42"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.366124 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c62ce10b-8219-4394-b6c4-4514b202da42" (UID: "c62ce10b-8219-4394-b6c4-4514b202da42"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.398584 4796 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.398652 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.398672 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x886v\" (UniqueName: \"kubernetes.io/projected/c62ce10b-8219-4394-b6c4-4514b202da42-kube-api-access-x886v\") on node \"crc\" DevicePath \"\"" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.398689 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62ce10b-8219-4394-b6c4-4514b202da42-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.894506 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" event={"ID":"c62ce10b-8219-4394-b6c4-4514b202da42","Type":"ContainerDied","Data":"ebc41d7bb478b68565e0571f458c1dbde517c62af08751cfe363a329d51baa98"} Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.894918 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebc41d7bb478b68565e0571f458c1dbde517c62af08751cfe363a329d51baa98" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.894572 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.968173 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc"] Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.973106 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="registry-server" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.973224 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="registry-server" Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.973321 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="extract-utilities" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.973413 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="extract-utilities" Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.973506 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="extract-utilities" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.973586 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="extract-utilities" Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.973682 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="registry-server" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.973765 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="registry-server" Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.973856 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62ce10b-8219-4394-b6c4-4514b202da42" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.973934 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62ce10b-8219-4394-b6c4-4514b202da42" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.974064 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="extract-content" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.974153 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="extract-content" Sep 30 16:36:25 crc kubenswrapper[4796]: E0930 16:36:25.974270 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="extract-content" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.974357 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="extract-content" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.974672 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62ce10b-8219-4394-b6c4-4514b202da42" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.974778 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ca5376-a635-4ce5-9089-e478f4534e88" containerName="registry-server" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.974886 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="392ebd6c-1211-4357-a605-ec8c5fe76337" containerName="registry-server" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.975734 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.979474 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.979900 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.979953 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.982507 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc"] Sep 30 16:36:25 crc kubenswrapper[4796]: I0930 16:36:25.982948 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.013359 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhhlk\" (UniqueName: \"kubernetes.io/projected/2d41515e-6a93-42d8-887c-c4f0f748bf3f-kube-api-access-xhhlk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.013538 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.013621 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.114806 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhhlk\" (UniqueName: \"kubernetes.io/projected/2d41515e-6a93-42d8-887c-c4f0f748bf3f-kube-api-access-xhhlk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.115070 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.116215 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.120472 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.120642 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.135287 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhhlk\" (UniqueName: \"kubernetes.io/projected/2d41515e-6a93-42d8-887c-c4f0f748bf3f-kube-api-access-xhhlk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.310753 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.908594 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc"] Sep 30 16:36:26 crc kubenswrapper[4796]: I0930 16:36:26.915210 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:36:27 crc kubenswrapper[4796]: I0930 16:36:27.920408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" event={"ID":"2d41515e-6a93-42d8-887c-c4f0f748bf3f","Type":"ContainerStarted","Data":"6aac6f843b705f3bbdfd6a47a6efc909649ddddac4fc2641dcfeb6d7ce27e343"} Sep 30 16:36:28 crc kubenswrapper[4796]: I0930 16:36:28.929297 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" event={"ID":"2d41515e-6a93-42d8-887c-c4f0f748bf3f","Type":"ContainerStarted","Data":"9274fae275fbcd9198f3cf7f5aa8ff11f14c92fedcf17bf10dd424198ab29b13"} Sep 30 16:36:28 crc kubenswrapper[4796]: I0930 16:36:28.951169 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" podStartSLOduration=2.761048271 podStartE2EDuration="3.951151571s" podCreationTimestamp="2025-09-30 16:36:25 +0000 UTC" firstStartedPulling="2025-09-30 16:36:26.914708853 +0000 UTC m=+1478.927987420" lastFinishedPulling="2025-09-30 16:36:28.104812183 +0000 UTC m=+1480.118090720" observedRunningTime="2025-09-30 16:36:28.946745474 +0000 UTC m=+1480.960024051" watchObservedRunningTime="2025-09-30 16:36:28.951151571 +0000 UTC m=+1480.964430098" Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.097097 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.097340 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.097392 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.098079 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.098124 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" gracePeriod=600 Sep 30 16:36:31 crc kubenswrapper[4796]: E0930 16:36:31.233159 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.962708 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" exitCode=0 Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.962760 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2"} Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.962802 4796 scope.go:117] "RemoveContainer" containerID="7b7ed0c28fbd614f1b48f9a4c2dd1b94dd52891d55126b94c8724c34b2b684f2" Sep 30 16:36:31 crc kubenswrapper[4796]: I0930 16:36:31.963474 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:36:31 crc kubenswrapper[4796]: E0930 16:36:31.963770 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:36:43 crc kubenswrapper[4796]: I0930 16:36:43.733582 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:36:43 crc kubenswrapper[4796]: E0930 16:36:43.734275 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:36:55 crc kubenswrapper[4796]: I0930 16:36:55.734444 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:36:55 crc kubenswrapper[4796]: E0930 16:36:55.735398 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:37:07 crc kubenswrapper[4796]: I0930 16:37:07.733362 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:37:07 crc kubenswrapper[4796]: E0930 16:37:07.734005 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:37:20 crc kubenswrapper[4796]: I0930 16:37:20.733456 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:37:20 crc kubenswrapper[4796]: E0930 16:37:20.734267 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:37:25 crc kubenswrapper[4796]: I0930 16:37:25.045424 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-c8s6r"] Sep 30 16:37:25 crc kubenswrapper[4796]: I0930 16:37:25.054776 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-c8s6r"] Sep 30 16:37:26 crc kubenswrapper[4796]: I0930 16:37:26.747313 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b" path="/var/lib/kubelet/pods/b6d7e8fa-83e5-4a0d-88e0-31594fe5c49b/volumes" Sep 30 16:37:30 crc kubenswrapper[4796]: I0930 16:37:30.048836 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-q2wwg"] Sep 30 16:37:30 crc kubenswrapper[4796]: I0930 16:37:30.059309 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-l49w6"] Sep 30 16:37:30 crc kubenswrapper[4796]: I0930 16:37:30.075854 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-q2wwg"] Sep 30 16:37:30 crc kubenswrapper[4796]: I0930 16:37:30.083147 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-l49w6"] Sep 30 16:37:30 crc kubenswrapper[4796]: I0930 16:37:30.755669 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06de8b69-fc4c-48b6-8fd6-6862a82b5620" path="/var/lib/kubelet/pods/06de8b69-fc4c-48b6-8fd6-6862a82b5620/volumes" Sep 30 16:37:30 crc kubenswrapper[4796]: I0930 16:37:30.756282 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed" path="/var/lib/kubelet/pods/cafd3c1f-17cb-426c-8dbe-f0f8dc0bc5ed/volumes" Sep 30 16:37:34 crc kubenswrapper[4796]: I0930 16:37:34.733515 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:37:34 crc kubenswrapper[4796]: E0930 16:37:34.734366 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:37:35 crc kubenswrapper[4796]: I0930 16:37:35.048302 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-dea9-account-create-t69m5"] Sep 30 16:37:35 crc kubenswrapper[4796]: I0930 16:37:35.065195 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-dea9-account-create-t69m5"] Sep 30 16:37:36 crc kubenswrapper[4796]: I0930 16:37:36.752038 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4ec0db4-b42d-4e72-86ce-e8de9acbe63f" path="/var/lib/kubelet/pods/f4ec0db4-b42d-4e72-86ce-e8de9acbe63f/volumes" Sep 30 16:37:40 crc kubenswrapper[4796]: I0930 16:37:40.034028 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8478-account-create-f9gfb"] Sep 30 16:37:40 crc kubenswrapper[4796]: I0930 16:37:40.045744 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8478-account-create-f9gfb"] Sep 30 16:37:40 crc kubenswrapper[4796]: I0930 16:37:40.748004 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e52c40-093e-4b19-aa48-ce91889790aa" path="/var/lib/kubelet/pods/59e52c40-093e-4b19-aa48-ce91889790aa/volumes" Sep 30 16:37:41 crc kubenswrapper[4796]: I0930 16:37:41.033310 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-995c-account-create-bprn8"] Sep 30 16:37:41 crc kubenswrapper[4796]: I0930 16:37:41.045654 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-995c-account-create-bprn8"] Sep 30 16:37:42 crc kubenswrapper[4796]: I0930 16:37:42.746480 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b203ad29-5d4a-4cb6-9fca-ea89ec8576f3" path="/var/lib/kubelet/pods/b203ad29-5d4a-4cb6-9fca-ea89ec8576f3/volumes" Sep 30 16:37:49 crc kubenswrapper[4796]: I0930 16:37:49.733538 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:37:49 crc kubenswrapper[4796]: E0930 16:37:49.734470 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.265623 4796 scope.go:117] "RemoveContainer" containerID="988ba621f63a1e38a73f1c81041a63a6b256cc28ce5cc5466f8ba555d09e7697" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.300143 4796 scope.go:117] "RemoveContainer" containerID="b9217eed3cb3bbd9aafc49fc5bfe297476ae6026c8fbae38548b840469436ef7" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.344566 4796 scope.go:117] "RemoveContainer" containerID="ad143d078d8ed653b598e264505cbf207b688cc9c70df7af71005abef2d4d7ca" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.392778 4796 scope.go:117] "RemoveContainer" containerID="016b0336b39e8f5b2712aec39f30b4e8ce36ae2a342735604b6b2804081f69cb" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.419449 4796 scope.go:117] "RemoveContainer" containerID="b3e0c4f6ec320b26192f04c26c7e6b2375db967b7d2e999c175ca55bba670a50" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.465816 4796 scope.go:117] "RemoveContainer" containerID="0b6ca28cd7dab8dd7c3b3be9c24dab13c01715679360a50c2e672f5a2b0ceb05" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.516797 4796 scope.go:117] "RemoveContainer" containerID="89d923d17b385bad1bcb3c5c70e5688be4bd3fe4a17652dc29efb54a8208e47c" Sep 30 16:37:53 crc kubenswrapper[4796]: I0930 16:37:53.559923 4796 scope.go:117] "RemoveContainer" containerID="3c0cd8a3cc475176ea5ea6cd2f86893bb0cc893082c69e0999f53eb0bfa60e55" Sep 30 16:37:57 crc kubenswrapper[4796]: I0930 16:37:57.040501 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tfr8r"] Sep 30 16:37:57 crc kubenswrapper[4796]: I0930 16:37:57.049276 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tfr8r"] Sep 30 16:37:58 crc kubenswrapper[4796]: I0930 16:37:58.751098 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="377e237f-78ad-4140-b988-690a2319c976" path="/var/lib/kubelet/pods/377e237f-78ad-4140-b988-690a2319c976/volumes" Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.028474 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9mss5"] Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.035512 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-942bk"] Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.043817 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-6bm8h"] Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.053547 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-942bk"] Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.064289 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-6bm8h"] Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.071265 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9mss5"] Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.747350 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c42976-648c-449e-9eb7-e876a4c3e67a" path="/var/lib/kubelet/pods/10c42976-648c-449e-9eb7-e876a4c3e67a/volumes" Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.749274 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3dec6a-f05e-4320-a625-d89fdb79f2d4" path="/var/lib/kubelet/pods/cc3dec6a-f05e-4320-a625-d89fdb79f2d4/volumes" Sep 30 16:38:00 crc kubenswrapper[4796]: I0930 16:38:00.750569 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe5beba9-e80d-45fd-90c8-72c7dbf2330f" path="/var/lib/kubelet/pods/fe5beba9-e80d-45fd-90c8-72c7dbf2330f/volumes" Sep 30 16:38:03 crc kubenswrapper[4796]: I0930 16:38:03.732972 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:38:03 crc kubenswrapper[4796]: E0930 16:38:03.733500 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:38:05 crc kubenswrapper[4796]: I0930 16:38:05.031404 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-6d4xh"] Sep 30 16:38:05 crc kubenswrapper[4796]: I0930 16:38:05.039747 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-6d4xh"] Sep 30 16:38:06 crc kubenswrapper[4796]: I0930 16:38:06.755100 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4a5eee5-b362-4420-9cb1-29abf8a46d3a" path="/var/lib/kubelet/pods/a4a5eee5-b362-4420-9cb1-29abf8a46d3a/volumes" Sep 30 16:38:07 crc kubenswrapper[4796]: I0930 16:38:07.029897 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-96a6-account-create-xfrx4"] Sep 30 16:38:07 crc kubenswrapper[4796]: I0930 16:38:07.041881 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-96a6-account-create-xfrx4"] Sep 30 16:38:08 crc kubenswrapper[4796]: I0930 16:38:08.758741 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b39aff1c-79f0-4747-a111-7c5ac3a7dc3f" path="/var/lib/kubelet/pods/b39aff1c-79f0-4747-a111-7c5ac3a7dc3f/volumes" Sep 30 16:38:17 crc kubenswrapper[4796]: I0930 16:38:17.733608 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:38:17 crc kubenswrapper[4796]: E0930 16:38:17.734379 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:38:18 crc kubenswrapper[4796]: I0930 16:38:18.044424 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b5da-account-create-l97vx"] Sep 30 16:38:18 crc kubenswrapper[4796]: I0930 16:38:18.053575 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b5da-account-create-l97vx"] Sep 30 16:38:18 crc kubenswrapper[4796]: I0930 16:38:18.743384 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a102b3-d988-4fed-837c-9217be9c9c2b" path="/var/lib/kubelet/pods/18a102b3-d988-4fed-837c-9217be9c9c2b/volumes" Sep 30 16:38:19 crc kubenswrapper[4796]: I0930 16:38:19.038561 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xp6sc"] Sep 30 16:38:19 crc kubenswrapper[4796]: I0930 16:38:19.050411 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xp6sc"] Sep 30 16:38:19 crc kubenswrapper[4796]: I0930 16:38:19.061361 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c1c4-account-create-52985"] Sep 30 16:38:19 crc kubenswrapper[4796]: I0930 16:38:19.072073 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c1c4-account-create-52985"] Sep 30 16:38:20 crc kubenswrapper[4796]: I0930 16:38:20.743514 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164" path="/var/lib/kubelet/pods/3c0837f7-80cf-4c8d-bf22-5c2d0dc1d164/volumes" Sep 30 16:38:20 crc kubenswrapper[4796]: I0930 16:38:20.744427 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604035fc-9731-47c4-b962-329eb8703d89" path="/var/lib/kubelet/pods/604035fc-9731-47c4-b962-329eb8703d89/volumes" Sep 30 16:38:27 crc kubenswrapper[4796]: I0930 16:38:27.039411 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dnmm6"] Sep 30 16:38:27 crc kubenswrapper[4796]: I0930 16:38:27.049236 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dnmm6"] Sep 30 16:38:28 crc kubenswrapper[4796]: I0930 16:38:28.747111 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b07f5969-8f8f-4976-99d8-a484adec2076" path="/var/lib/kubelet/pods/b07f5969-8f8f-4976-99d8-a484adec2076/volumes" Sep 30 16:38:30 crc kubenswrapper[4796]: I0930 16:38:30.734408 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:38:30 crc kubenswrapper[4796]: E0930 16:38:30.735085 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:38:39 crc kubenswrapper[4796]: I0930 16:38:39.383403 4796 generic.go:334] "Generic (PLEG): container finished" podID="2d41515e-6a93-42d8-887c-c4f0f748bf3f" containerID="9274fae275fbcd9198f3cf7f5aa8ff11f14c92fedcf17bf10dd424198ab29b13" exitCode=0 Sep 30 16:38:39 crc kubenswrapper[4796]: I0930 16:38:39.383506 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" event={"ID":"2d41515e-6a93-42d8-887c-c4f0f748bf3f","Type":"ContainerDied","Data":"9274fae275fbcd9198f3cf7f5aa8ff11f14c92fedcf17bf10dd424198ab29b13"} Sep 30 16:38:40 crc kubenswrapper[4796]: I0930 16:38:40.863530 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:38:40 crc kubenswrapper[4796]: I0930 16:38:40.991936 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhhlk\" (UniqueName: \"kubernetes.io/projected/2d41515e-6a93-42d8-887c-c4f0f748bf3f-kube-api-access-xhhlk\") pod \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " Sep 30 16:38:40 crc kubenswrapper[4796]: I0930 16:38:40.992141 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-ssh-key\") pod \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " Sep 30 16:38:40 crc kubenswrapper[4796]: I0930 16:38:40.992233 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-inventory\") pod \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\" (UID: \"2d41515e-6a93-42d8-887c-c4f0f748bf3f\") " Sep 30 16:38:40 crc kubenswrapper[4796]: I0930 16:38:40.998365 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d41515e-6a93-42d8-887c-c4f0f748bf3f-kube-api-access-xhhlk" (OuterVolumeSpecName: "kube-api-access-xhhlk") pod "2d41515e-6a93-42d8-887c-c4f0f748bf3f" (UID: "2d41515e-6a93-42d8-887c-c4f0f748bf3f"). InnerVolumeSpecName "kube-api-access-xhhlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.024223 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2d41515e-6a93-42d8-887c-c4f0f748bf3f" (UID: "2d41515e-6a93-42d8-887c-c4f0f748bf3f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.043104 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-inventory" (OuterVolumeSpecName: "inventory") pod "2d41515e-6a93-42d8-887c-c4f0f748bf3f" (UID: "2d41515e-6a93-42d8-887c-c4f0f748bf3f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.095604 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.095652 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d41515e-6a93-42d8-887c-c4f0f748bf3f-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.095665 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhhlk\" (UniqueName: \"kubernetes.io/projected/2d41515e-6a93-42d8-887c-c4f0f748bf3f-kube-api-access-xhhlk\") on node \"crc\" DevicePath \"\"" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.405733 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" event={"ID":"2d41515e-6a93-42d8-887c-c4f0f748bf3f","Type":"ContainerDied","Data":"6aac6f843b705f3bbdfd6a47a6efc909649ddddac4fc2641dcfeb6d7ce27e343"} Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.405773 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6aac6f843b705f3bbdfd6a47a6efc909649ddddac4fc2641dcfeb6d7ce27e343" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.405824 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.508437 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw"] Sep 30 16:38:41 crc kubenswrapper[4796]: E0930 16:38:41.508907 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d41515e-6a93-42d8-887c-c4f0f748bf3f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.508932 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d41515e-6a93-42d8-887c-c4f0f748bf3f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.509192 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d41515e-6a93-42d8-887c-c4f0f748bf3f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.509911 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.512409 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.513494 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.514432 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.514638 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.528537 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw"] Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.602872 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47mr6\" (UniqueName: \"kubernetes.io/projected/6ba38983-af8e-47bc-9081-7404ef0dfefc-kube-api-access-47mr6\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.602970 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.603122 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.704565 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47mr6\" (UniqueName: \"kubernetes.io/projected/6ba38983-af8e-47bc-9081-7404ef0dfefc-kube-api-access-47mr6\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.705102 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.705161 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.708919 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.709006 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.719964 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47mr6\" (UniqueName: \"kubernetes.io/projected/6ba38983-af8e-47bc-9081-7404ef0dfefc-kube-api-access-47mr6\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:41 crc kubenswrapper[4796]: I0930 16:38:41.825590 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:38:42 crc kubenswrapper[4796]: I0930 16:38:42.403548 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw"] Sep 30 16:38:43 crc kubenswrapper[4796]: I0930 16:38:43.428071 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" event={"ID":"6ba38983-af8e-47bc-9081-7404ef0dfefc","Type":"ContainerStarted","Data":"6744e790d7f3954cad3d9ba7d451fbc0007a1df966c8b83d2f608ddd5a588111"} Sep 30 16:38:43 crc kubenswrapper[4796]: I0930 16:38:43.428455 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" event={"ID":"6ba38983-af8e-47bc-9081-7404ef0dfefc","Type":"ContainerStarted","Data":"409a89b38bed1c66f72666ed6866ceed90f9325a7b06e1ce977a4771837cb7da"} Sep 30 16:38:43 crc kubenswrapper[4796]: I0930 16:38:43.452906 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" podStartSLOduration=1.8174265790000002 podStartE2EDuration="2.452888557s" podCreationTimestamp="2025-09-30 16:38:41 +0000 UTC" firstStartedPulling="2025-09-30 16:38:42.415065841 +0000 UTC m=+1614.428344368" lastFinishedPulling="2025-09-30 16:38:43.050527799 +0000 UTC m=+1615.063806346" observedRunningTime="2025-09-30 16:38:43.442523267 +0000 UTC m=+1615.455801794" watchObservedRunningTime="2025-09-30 16:38:43.452888557 +0000 UTC m=+1615.466167084" Sep 30 16:38:44 crc kubenswrapper[4796]: I0930 16:38:44.734057 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:38:44 crc kubenswrapper[4796]: E0930 16:38:44.734491 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:38:47 crc kubenswrapper[4796]: I0930 16:38:47.045588 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-rsljj"] Sep 30 16:38:47 crc kubenswrapper[4796]: I0930 16:38:47.052764 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-rsljj"] Sep 30 16:38:48 crc kubenswrapper[4796]: I0930 16:38:48.753103 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15717b4f-8c0e-4105-a2da-1fb7c502a4cc" path="/var/lib/kubelet/pods/15717b4f-8c0e-4105-a2da-1fb7c502a4cc/volumes" Sep 30 16:38:53 crc kubenswrapper[4796]: I0930 16:38:53.727840 4796 scope.go:117] "RemoveContainer" containerID="6fe006240ef270bbef55215ea4c3cec1191ff6021a600a638b1de4a76384da33" Sep 30 16:38:53 crc kubenswrapper[4796]: I0930 16:38:53.768008 4796 scope.go:117] "RemoveContainer" containerID="cdddeec154684c68a56bb45eca5755fac783ede66ea17ac56b59a529131257eb" Sep 30 16:38:53 crc kubenswrapper[4796]: I0930 16:38:53.832011 4796 scope.go:117] "RemoveContainer" containerID="58bf3680142c54a61523f1fc71b893bc28f4670b5c4c2b2d246c1a717f1336d9" Sep 30 16:38:53 crc kubenswrapper[4796]: I0930 16:38:53.878903 4796 scope.go:117] "RemoveContainer" containerID="4a9955c3325e12f6a19b92fbb7386c490103157c5c5200a90824b06993dd16a4" Sep 30 16:38:53 crc kubenswrapper[4796]: I0930 16:38:53.934967 4796 scope.go:117] "RemoveContainer" containerID="884af2dce8fbb96f9fc0a188df28b12eec1735e11851391a7b8c5f03fff8e547" Sep 30 16:38:53 crc kubenswrapper[4796]: I0930 16:38:53.965492 4796 scope.go:117] "RemoveContainer" containerID="65edba31166889b96c3113729099421540f1f89f256dd4f60863d8db753ec64f" Sep 30 16:38:54 crc kubenswrapper[4796]: I0930 16:38:54.005463 4796 scope.go:117] "RemoveContainer" containerID="feb7ef30d37b7763c79a804fab949915cc8a68ea9292fc9399bc71c8e95620e5" Sep 30 16:38:54 crc kubenswrapper[4796]: I0930 16:38:54.023921 4796 scope.go:117] "RemoveContainer" containerID="1e1f238a04e2217ed3a436ae804da29e89f49e193166068df63f3d9af4b30cc0" Sep 30 16:38:54 crc kubenswrapper[4796]: I0930 16:38:54.058249 4796 scope.go:117] "RemoveContainer" containerID="957b10dc4d6aa188ce92c1cb6ebdf63d73b274b07cd034d81c222699103b7543" Sep 30 16:38:54 crc kubenswrapper[4796]: I0930 16:38:54.093504 4796 scope.go:117] "RemoveContainer" containerID="7961f219c18f25894dfe345b292a8d4746ac437570709ea4cbcc86881889a7e1" Sep 30 16:38:54 crc kubenswrapper[4796]: I0930 16:38:54.135048 4796 scope.go:117] "RemoveContainer" containerID="98b3b3af45327799ac7ad8d353f6163d748490a9e8f7e6181a055df49f6a6b7f" Sep 30 16:38:55 crc kubenswrapper[4796]: I0930 16:38:55.029115 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-dxs8h"] Sep 30 16:38:55 crc kubenswrapper[4796]: I0930 16:38:55.036849 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-dxs8h"] Sep 30 16:38:56 crc kubenswrapper[4796]: I0930 16:38:56.751767 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27218284-a0ab-4297-8d33-d9abc4e1549f" path="/var/lib/kubelet/pods/27218284-a0ab-4297-8d33-d9abc4e1549f/volumes" Sep 30 16:38:59 crc kubenswrapper[4796]: I0930 16:38:59.733373 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:38:59 crc kubenswrapper[4796]: E0930 16:38:59.733888 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:39:00 crc kubenswrapper[4796]: I0930 16:39:00.040969 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-rtdjm"] Sep 30 16:39:00 crc kubenswrapper[4796]: I0930 16:39:00.050176 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-rtdjm"] Sep 30 16:39:00 crc kubenswrapper[4796]: I0930 16:39:00.754366 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d5f8250-66ae-4368-b6d6-7d0bea34d616" path="/var/lib/kubelet/pods/4d5f8250-66ae-4368-b6d6-7d0bea34d616/volumes" Sep 30 16:39:10 crc kubenswrapper[4796]: I0930 16:39:10.733834 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:39:10 crc kubenswrapper[4796]: E0930 16:39:10.734769 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:39:23 crc kubenswrapper[4796]: I0930 16:39:23.733309 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:39:23 crc kubenswrapper[4796]: E0930 16:39:23.734509 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:39:31 crc kubenswrapper[4796]: I0930 16:39:31.046135 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4jh9b"] Sep 30 16:39:31 crc kubenswrapper[4796]: I0930 16:39:31.058266 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4jh9b"] Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.040828 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jx95c"] Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.052586 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-494pn"] Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.060342 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jx95c"] Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.067399 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-494pn"] Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.746511 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c1cb023-494f-4318-8b3b-bc7d67977219" path="/var/lib/kubelet/pods/5c1cb023-494f-4318-8b3b-bc7d67977219/volumes" Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.747635 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c70e4cae-e24d-418e-8f54-e377aa2708d3" path="/var/lib/kubelet/pods/c70e4cae-e24d-418e-8f54-e377aa2708d3/volumes" Sep 30 16:39:32 crc kubenswrapper[4796]: I0930 16:39:32.748532 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df2a9cd6-ca55-412d-9862-c4a59b44fa54" path="/var/lib/kubelet/pods/df2a9cd6-ca55-412d-9862-c4a59b44fa54/volumes" Sep 30 16:39:38 crc kubenswrapper[4796]: I0930 16:39:38.745483 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:39:38 crc kubenswrapper[4796]: E0930 16:39:38.746659 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:39:41 crc kubenswrapper[4796]: I0930 16:39:41.034772 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1fcc-account-create-ngpsd"] Sep 30 16:39:41 crc kubenswrapper[4796]: I0930 16:39:41.048723 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1fcc-account-create-ngpsd"] Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.032603 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6979-account-create-nlhts"] Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.043976 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5721-account-create-6k8xp"] Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.060415 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6979-account-create-nlhts"] Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.067406 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5721-account-create-6k8xp"] Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.746550 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087eaf83-481b-4b24-a3e8-3a674de51028" path="/var/lib/kubelet/pods/087eaf83-481b-4b24-a3e8-3a674de51028/volumes" Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.747215 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16e579e1-4e1c-4b03-9a22-af7399019ac4" path="/var/lib/kubelet/pods/16e579e1-4e1c-4b03-9a22-af7399019ac4/volumes" Sep 30 16:39:42 crc kubenswrapper[4796]: I0930 16:39:42.747748 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fee951a-80a3-4086-a18a-7121e4402a07" path="/var/lib/kubelet/pods/7fee951a-80a3-4086-a18a-7121e4402a07/volumes" Sep 30 16:39:52 crc kubenswrapper[4796]: I0930 16:39:52.734164 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:39:52 crc kubenswrapper[4796]: E0930 16:39:52.735259 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.360032 4796 scope.go:117] "RemoveContainer" containerID="95b0f331fcab6b992b9f0b3330f7d4889971201384740512d8c4870b3f20cde1" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.420618 4796 scope.go:117] "RemoveContainer" containerID="86ef7588e5d2a520dc64bd5980f777c8575e3983691cfdee37470da0f88dd78a" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.453970 4796 scope.go:117] "RemoveContainer" containerID="7237daed360dc265a07da547ed5916f32dce47c45c71b2d5c28c5b2e59df4ff8" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.530532 4796 scope.go:117] "RemoveContainer" containerID="b92956c52884a939123cf26ef0ee6bee755252ed7edb05ee5551d1fa800820df" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.556501 4796 scope.go:117] "RemoveContainer" containerID="a611bccd55e2ffce8221c0381c93b8542ef5e03c5c7b9496d4ab9a9f2e2c453e" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.594071 4796 scope.go:117] "RemoveContainer" containerID="7f110dd7e4ecdf1f0e79d45a53b07abdcb75ca66393e280322686d144aefbc8f" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.661335 4796 scope.go:117] "RemoveContainer" containerID="b87b361f05d2973a3fdd3eedb2e30570cfc502408e856dc1877050a44cfd2eba" Sep 30 16:39:54 crc kubenswrapper[4796]: I0930 16:39:54.691852 4796 scope.go:117] "RemoveContainer" containerID="98c13e2cf727be9d106a4804a2e659d109f6f410e018070c6d6f97def1af8a85" Sep 30 16:40:04 crc kubenswrapper[4796]: I0930 16:40:04.069715 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fvd2d"] Sep 30 16:40:04 crc kubenswrapper[4796]: I0930 16:40:04.085336 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fvd2d"] Sep 30 16:40:04 crc kubenswrapper[4796]: I0930 16:40:04.744807 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e4ce38d-a600-48e0-8178-676b1bbd17c1" path="/var/lib/kubelet/pods/1e4ce38d-a600-48e0-8178-676b1bbd17c1/volumes" Sep 30 16:40:06 crc kubenswrapper[4796]: I0930 16:40:06.330040 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ba38983-af8e-47bc-9081-7404ef0dfefc" containerID="6744e790d7f3954cad3d9ba7d451fbc0007a1df966c8b83d2f608ddd5a588111" exitCode=0 Sep 30 16:40:06 crc kubenswrapper[4796]: I0930 16:40:06.330150 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" event={"ID":"6ba38983-af8e-47bc-9081-7404ef0dfefc","Type":"ContainerDied","Data":"6744e790d7f3954cad3d9ba7d451fbc0007a1df966c8b83d2f608ddd5a588111"} Sep 30 16:40:06 crc kubenswrapper[4796]: I0930 16:40:06.733883 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:40:06 crc kubenswrapper[4796]: E0930 16:40:06.734845 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.827993 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.939490 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-inventory\") pod \"6ba38983-af8e-47bc-9081-7404ef0dfefc\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.939559 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47mr6\" (UniqueName: \"kubernetes.io/projected/6ba38983-af8e-47bc-9081-7404ef0dfefc-kube-api-access-47mr6\") pod \"6ba38983-af8e-47bc-9081-7404ef0dfefc\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.939816 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-ssh-key\") pod \"6ba38983-af8e-47bc-9081-7404ef0dfefc\" (UID: \"6ba38983-af8e-47bc-9081-7404ef0dfefc\") " Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.956016 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba38983-af8e-47bc-9081-7404ef0dfefc-kube-api-access-47mr6" (OuterVolumeSpecName: "kube-api-access-47mr6") pod "6ba38983-af8e-47bc-9081-7404ef0dfefc" (UID: "6ba38983-af8e-47bc-9081-7404ef0dfefc"). InnerVolumeSpecName "kube-api-access-47mr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.970838 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-inventory" (OuterVolumeSpecName: "inventory") pod "6ba38983-af8e-47bc-9081-7404ef0dfefc" (UID: "6ba38983-af8e-47bc-9081-7404ef0dfefc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:40:07 crc kubenswrapper[4796]: I0930 16:40:07.978198 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ba38983-af8e-47bc-9081-7404ef0dfefc" (UID: "6ba38983-af8e-47bc-9081-7404ef0dfefc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.042233 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.042275 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ba38983-af8e-47bc-9081-7404ef0dfefc-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.042294 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47mr6\" (UniqueName: \"kubernetes.io/projected/6ba38983-af8e-47bc-9081-7404ef0dfefc-kube-api-access-47mr6\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.356331 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" event={"ID":"6ba38983-af8e-47bc-9081-7404ef0dfefc","Type":"ContainerDied","Data":"409a89b38bed1c66f72666ed6866ceed90f9325a7b06e1ce977a4771837cb7da"} Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.356377 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="409a89b38bed1c66f72666ed6866ceed90f9325a7b06e1ce977a4771837cb7da" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.356442 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.468282 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv"] Sep 30 16:40:08 crc kubenswrapper[4796]: E0930 16:40:08.468915 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba38983-af8e-47bc-9081-7404ef0dfefc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.469087 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba38983-af8e-47bc-9081-7404ef0dfefc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.469437 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba38983-af8e-47bc-9081-7404ef0dfefc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.470350 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.473029 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.473385 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.473513 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.482851 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv"] Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.482953 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.654539 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.654970 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvpkc\" (UniqueName: \"kubernetes.io/projected/2e9c9f25-af36-4281-b747-d03a289f22f5-kube-api-access-hvpkc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.655374 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.757364 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.757604 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.757656 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvpkc\" (UniqueName: \"kubernetes.io/projected/2e9c9f25-af36-4281-b747-d03a289f22f5-kube-api-access-hvpkc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.762373 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.763927 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.776398 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvpkc\" (UniqueName: \"kubernetes.io/projected/2e9c9f25-af36-4281-b747-d03a289f22f5-kube-api-access-hvpkc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b77pv\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:08 crc kubenswrapper[4796]: I0930 16:40:08.787373 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:09 crc kubenswrapper[4796]: I0930 16:40:09.346602 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv"] Sep 30 16:40:09 crc kubenswrapper[4796]: I0930 16:40:09.366815 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" event={"ID":"2e9c9f25-af36-4281-b747-d03a289f22f5","Type":"ContainerStarted","Data":"e99be31dd62a3aee6c1460b3c5d928fddea66107ca1e5c4952d80d68d31b8f87"} Sep 30 16:40:10 crc kubenswrapper[4796]: I0930 16:40:10.386793 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" event={"ID":"2e9c9f25-af36-4281-b747-d03a289f22f5","Type":"ContainerStarted","Data":"3181df4d7fb4ec2ee2eb20ecdce75ef610c607fa85b037521e4bdda9ad616c7f"} Sep 30 16:40:10 crc kubenswrapper[4796]: I0930 16:40:10.415043 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" podStartSLOduration=1.881537234 podStartE2EDuration="2.414960919s" podCreationTimestamp="2025-09-30 16:40:08 +0000 UTC" firstStartedPulling="2025-09-30 16:40:09.35618115 +0000 UTC m=+1701.369459677" lastFinishedPulling="2025-09-30 16:40:09.889604805 +0000 UTC m=+1701.902883362" observedRunningTime="2025-09-30 16:40:10.414415383 +0000 UTC m=+1702.427693930" watchObservedRunningTime="2025-09-30 16:40:10.414960919 +0000 UTC m=+1702.428239446" Sep 30 16:40:15 crc kubenswrapper[4796]: I0930 16:40:15.452332 4796 generic.go:334] "Generic (PLEG): container finished" podID="2e9c9f25-af36-4281-b747-d03a289f22f5" containerID="3181df4d7fb4ec2ee2eb20ecdce75ef610c607fa85b037521e4bdda9ad616c7f" exitCode=0 Sep 30 16:40:15 crc kubenswrapper[4796]: I0930 16:40:15.452443 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" event={"ID":"2e9c9f25-af36-4281-b747-d03a289f22f5","Type":"ContainerDied","Data":"3181df4d7fb4ec2ee2eb20ecdce75ef610c607fa85b037521e4bdda9ad616c7f"} Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.909402 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.941751 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvpkc\" (UniqueName: \"kubernetes.io/projected/2e9c9f25-af36-4281-b747-d03a289f22f5-kube-api-access-hvpkc\") pod \"2e9c9f25-af36-4281-b747-d03a289f22f5\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.941842 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-inventory\") pod \"2e9c9f25-af36-4281-b747-d03a289f22f5\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.942025 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-ssh-key\") pod \"2e9c9f25-af36-4281-b747-d03a289f22f5\" (UID: \"2e9c9f25-af36-4281-b747-d03a289f22f5\") " Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.955357 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e9c9f25-af36-4281-b747-d03a289f22f5-kube-api-access-hvpkc" (OuterVolumeSpecName: "kube-api-access-hvpkc") pod "2e9c9f25-af36-4281-b747-d03a289f22f5" (UID: "2e9c9f25-af36-4281-b747-d03a289f22f5"). InnerVolumeSpecName "kube-api-access-hvpkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.971196 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2e9c9f25-af36-4281-b747-d03a289f22f5" (UID: "2e9c9f25-af36-4281-b747-d03a289f22f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:40:16 crc kubenswrapper[4796]: I0930 16:40:16.984160 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-inventory" (OuterVolumeSpecName: "inventory") pod "2e9c9f25-af36-4281-b747-d03a289f22f5" (UID: "2e9c9f25-af36-4281-b747-d03a289f22f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.044089 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.044122 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvpkc\" (UniqueName: \"kubernetes.io/projected/2e9c9f25-af36-4281-b747-d03a289f22f5-kube-api-access-hvpkc\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.044133 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e9c9f25-af36-4281-b747-d03a289f22f5-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.481686 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" event={"ID":"2e9c9f25-af36-4281-b747-d03a289f22f5","Type":"ContainerDied","Data":"e99be31dd62a3aee6c1460b3c5d928fddea66107ca1e5c4952d80d68d31b8f87"} Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.481750 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e99be31dd62a3aee6c1460b3c5d928fddea66107ca1e5c4952d80d68d31b8f87" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.481818 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b77pv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.555278 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv"] Sep 30 16:40:17 crc kubenswrapper[4796]: E0930 16:40:17.555698 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e9c9f25-af36-4281-b747-d03a289f22f5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.555718 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e9c9f25-af36-4281-b747-d03a289f22f5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.555961 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e9c9f25-af36-4281-b747-d03a289f22f5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.556729 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.563683 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.563741 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.563778 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.564009 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.574701 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv"] Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.655822 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.655893 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4jtc\" (UniqueName: \"kubernetes.io/projected/743c014a-a242-4a7a-96e5-167a0d0928fb-kube-api-access-d4jtc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.656018 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.761376 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.761443 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4jtc\" (UniqueName: \"kubernetes.io/projected/743c014a-a242-4a7a-96e5-167a0d0928fb-kube-api-access-d4jtc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.761543 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.767527 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.783828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.783894 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4jtc\" (UniqueName: \"kubernetes.io/projected/743c014a-a242-4a7a-96e5-167a0d0928fb-kube-api-access-d4jtc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zm5lv\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:17 crc kubenswrapper[4796]: I0930 16:40:17.883195 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:40:18 crc kubenswrapper[4796]: I0930 16:40:18.417293 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv"] Sep 30 16:40:18 crc kubenswrapper[4796]: I0930 16:40:18.503654 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" event={"ID":"743c014a-a242-4a7a-96e5-167a0d0928fb","Type":"ContainerStarted","Data":"c518371290c2cdf582d72d1d1a74e3a2c468dc57d2511dd46f1ceec926497c2d"} Sep 30 16:40:19 crc kubenswrapper[4796]: I0930 16:40:19.518926 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" event={"ID":"743c014a-a242-4a7a-96e5-167a0d0928fb","Type":"ContainerStarted","Data":"7ed001b4236e4af6e20b2800c11037beeadd23bc12a419e3769d0ca9cc7e4d2a"} Sep 30 16:40:19 crc kubenswrapper[4796]: I0930 16:40:19.732856 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:40:19 crc kubenswrapper[4796]: E0930 16:40:19.733491 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:40:28 crc kubenswrapper[4796]: I0930 16:40:28.041410 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" podStartSLOduration=10.545690339 podStartE2EDuration="11.041389647s" podCreationTimestamp="2025-09-30 16:40:17 +0000 UTC" firstStartedPulling="2025-09-30 16:40:18.425575693 +0000 UTC m=+1710.438854230" lastFinishedPulling="2025-09-30 16:40:18.921275001 +0000 UTC m=+1710.934553538" observedRunningTime="2025-09-30 16:40:19.544824152 +0000 UTC m=+1711.558102739" watchObservedRunningTime="2025-09-30 16:40:28.041389647 +0000 UTC m=+1720.054668164" Sep 30 16:40:28 crc kubenswrapper[4796]: I0930 16:40:28.043202 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlmls"] Sep 30 16:40:28 crc kubenswrapper[4796]: I0930 16:40:28.049877 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlmls"] Sep 30 16:40:28 crc kubenswrapper[4796]: I0930 16:40:28.757152 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10a7487-db94-441a-856d-dff951ed56cf" path="/var/lib/kubelet/pods/f10a7487-db94-441a-856d-dff951ed56cf/volumes" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.227855 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dp9n7"] Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.230171 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.264714 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dp9n7"] Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.414183 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r2gf\" (UniqueName: \"kubernetes.io/projected/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-kube-api-access-7r2gf\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.414268 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-utilities\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.414733 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-catalog-content\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.517364 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r2gf\" (UniqueName: \"kubernetes.io/projected/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-kube-api-access-7r2gf\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.517532 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-utilities\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.517764 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-catalog-content\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.518184 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-utilities\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.518414 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-catalog-content\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.542906 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r2gf\" (UniqueName: \"kubernetes.io/projected/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-kube-api-access-7r2gf\") pod \"certified-operators-dp9n7\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:30 crc kubenswrapper[4796]: I0930 16:40:30.561038 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:31 crc kubenswrapper[4796]: I0930 16:40:31.094441 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dp9n7"] Sep 30 16:40:31 crc kubenswrapper[4796]: W0930 16:40:31.104855 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod970d4ab4_bebb_4fb0_b000_854bd7ed9e99.slice/crio-167ff12d5eede647681f3b49e8858b977c43ec9155d650ae7fe13234ebc2cf7d WatchSource:0}: Error finding container 167ff12d5eede647681f3b49e8858b977c43ec9155d650ae7fe13234ebc2cf7d: Status 404 returned error can't find the container with id 167ff12d5eede647681f3b49e8858b977c43ec9155d650ae7fe13234ebc2cf7d Sep 30 16:40:31 crc kubenswrapper[4796]: I0930 16:40:31.627468 4796 generic.go:334] "Generic (PLEG): container finished" podID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerID="5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51" exitCode=0 Sep 30 16:40:31 crc kubenswrapper[4796]: I0930 16:40:31.627569 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerDied","Data":"5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51"} Sep 30 16:40:31 crc kubenswrapper[4796]: I0930 16:40:31.627911 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerStarted","Data":"167ff12d5eede647681f3b49e8858b977c43ec9155d650ae7fe13234ebc2cf7d"} Sep 30 16:40:32 crc kubenswrapper[4796]: I0930 16:40:32.032784 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jpscz"] Sep 30 16:40:32 crc kubenswrapper[4796]: I0930 16:40:32.048649 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jpscz"] Sep 30 16:40:32 crc kubenswrapper[4796]: I0930 16:40:32.639540 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerStarted","Data":"cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240"} Sep 30 16:40:32 crc kubenswrapper[4796]: I0930 16:40:32.747180 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="800b5cc2-8806-4298-802a-7bef00b36a4c" path="/var/lib/kubelet/pods/800b5cc2-8806-4298-802a-7bef00b36a4c/volumes" Sep 30 16:40:33 crc kubenswrapper[4796]: I0930 16:40:33.654765 4796 generic.go:334] "Generic (PLEG): container finished" podID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerID="cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240" exitCode=0 Sep 30 16:40:33 crc kubenswrapper[4796]: I0930 16:40:33.654871 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerDied","Data":"cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240"} Sep 30 16:40:33 crc kubenswrapper[4796]: I0930 16:40:33.733266 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:40:33 crc kubenswrapper[4796]: E0930 16:40:33.733543 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:40:34 crc kubenswrapper[4796]: I0930 16:40:34.673968 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerStarted","Data":"8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9"} Sep 30 16:40:34 crc kubenswrapper[4796]: I0930 16:40:34.697430 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dp9n7" podStartSLOduration=2.261635952 podStartE2EDuration="4.697404454s" podCreationTimestamp="2025-09-30 16:40:30 +0000 UTC" firstStartedPulling="2025-09-30 16:40:31.629337958 +0000 UTC m=+1723.642616475" lastFinishedPulling="2025-09-30 16:40:34.06510643 +0000 UTC m=+1726.078384977" observedRunningTime="2025-09-30 16:40:34.693857801 +0000 UTC m=+1726.707136368" watchObservedRunningTime="2025-09-30 16:40:34.697404454 +0000 UTC m=+1726.710683011" Sep 30 16:40:40 crc kubenswrapper[4796]: I0930 16:40:40.561869 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:40 crc kubenswrapper[4796]: I0930 16:40:40.562609 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:40 crc kubenswrapper[4796]: I0930 16:40:40.626382 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:40 crc kubenswrapper[4796]: I0930 16:40:40.805909 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:41 crc kubenswrapper[4796]: I0930 16:40:41.208712 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dp9n7"] Sep 30 16:40:42 crc kubenswrapper[4796]: I0930 16:40:42.754470 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dp9n7" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="registry-server" containerID="cri-o://8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9" gracePeriod=2 Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.243380 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.384290 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-utilities\") pod \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.384443 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r2gf\" (UniqueName: \"kubernetes.io/projected/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-kube-api-access-7r2gf\") pod \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.384497 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-catalog-content\") pod \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\" (UID: \"970d4ab4-bebb-4fb0-b000-854bd7ed9e99\") " Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.385828 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-utilities" (OuterVolumeSpecName: "utilities") pod "970d4ab4-bebb-4fb0-b000-854bd7ed9e99" (UID: "970d4ab4-bebb-4fb0-b000-854bd7ed9e99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.390972 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-kube-api-access-7r2gf" (OuterVolumeSpecName: "kube-api-access-7r2gf") pod "970d4ab4-bebb-4fb0-b000-854bd7ed9e99" (UID: "970d4ab4-bebb-4fb0-b000-854bd7ed9e99"). InnerVolumeSpecName "kube-api-access-7r2gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.487443 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.487500 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r2gf\" (UniqueName: \"kubernetes.io/projected/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-kube-api-access-7r2gf\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.766351 4796 generic.go:334] "Generic (PLEG): container finished" podID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerID="8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9" exitCode=0 Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.766409 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerDied","Data":"8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9"} Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.766709 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dp9n7" event={"ID":"970d4ab4-bebb-4fb0-b000-854bd7ed9e99","Type":"ContainerDied","Data":"167ff12d5eede647681f3b49e8858b977c43ec9155d650ae7fe13234ebc2cf7d"} Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.766737 4796 scope.go:117] "RemoveContainer" containerID="8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.766429 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dp9n7" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.797467 4796 scope.go:117] "RemoveContainer" containerID="cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.830554 4796 scope.go:117] "RemoveContainer" containerID="5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.870657 4796 scope.go:117] "RemoveContainer" containerID="8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9" Sep 30 16:40:43 crc kubenswrapper[4796]: E0930 16:40:43.871166 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9\": container with ID starting with 8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9 not found: ID does not exist" containerID="8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.871197 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9"} err="failed to get container status \"8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9\": rpc error: code = NotFound desc = could not find container \"8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9\": container with ID starting with 8de001e47fec7bf09573cf39953ff604217520df998ff179d71880e2ca4a31e9 not found: ID does not exist" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.871219 4796 scope.go:117] "RemoveContainer" containerID="cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240" Sep 30 16:40:43 crc kubenswrapper[4796]: E0930 16:40:43.871662 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240\": container with ID starting with cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240 not found: ID does not exist" containerID="cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.871749 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240"} err="failed to get container status \"cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240\": rpc error: code = NotFound desc = could not find container \"cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240\": container with ID starting with cc8aa70cf586d8605523fbff9dc7a2b27745b1edefff45556ad3e02786b68240 not found: ID does not exist" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.871799 4796 scope.go:117] "RemoveContainer" containerID="5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51" Sep 30 16:40:43 crc kubenswrapper[4796]: E0930 16:40:43.872323 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51\": container with ID starting with 5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51 not found: ID does not exist" containerID="5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.872366 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51"} err="failed to get container status \"5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51\": rpc error: code = NotFound desc = could not find container \"5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51\": container with ID starting with 5e2d321d91a7c345c1c20782d6fad4b95ef612bea0c47be86ec0e2adcb02ce51 not found: ID does not exist" Sep 30 16:40:43 crc kubenswrapper[4796]: I0930 16:40:43.939411 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "970d4ab4-bebb-4fb0-b000-854bd7ed9e99" (UID: "970d4ab4-bebb-4fb0-b000-854bd7ed9e99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:40:44 crc kubenswrapper[4796]: I0930 16:40:44.002556 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d4ab4-bebb-4fb0-b000-854bd7ed9e99-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:40:44 crc kubenswrapper[4796]: I0930 16:40:44.102819 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dp9n7"] Sep 30 16:40:44 crc kubenswrapper[4796]: I0930 16:40:44.109814 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dp9n7"] Sep 30 16:40:44 crc kubenswrapper[4796]: I0930 16:40:44.748208 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" path="/var/lib/kubelet/pods/970d4ab4-bebb-4fb0-b000-854bd7ed9e99/volumes" Sep 30 16:40:47 crc kubenswrapper[4796]: I0930 16:40:47.733897 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:40:47 crc kubenswrapper[4796]: E0930 16:40:47.734821 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:40:54 crc kubenswrapper[4796]: I0930 16:40:54.902468 4796 scope.go:117] "RemoveContainer" containerID="c1eb20b2801e7239d54a8f0baee6f4a3f6b073d266e3fddc78bb29a987925897" Sep 30 16:40:54 crc kubenswrapper[4796]: I0930 16:40:54.970914 4796 scope.go:117] "RemoveContainer" containerID="d1e335a5c446dfe36da0709c7daadfd01ef39f4982921030c7fc75692148aa8e" Sep 30 16:40:55 crc kubenswrapper[4796]: I0930 16:40:55.038545 4796 scope.go:117] "RemoveContainer" containerID="a57aa13db2e426dfd044edd4989d291ba10b52a8d7e36eb6324e3600e091bc27" Sep 30 16:40:58 crc kubenswrapper[4796]: I0930 16:40:58.744856 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:40:58 crc kubenswrapper[4796]: E0930 16:40:58.745625 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:40:59 crc kubenswrapper[4796]: I0930 16:40:59.945664 4796 generic.go:334] "Generic (PLEG): container finished" podID="743c014a-a242-4a7a-96e5-167a0d0928fb" containerID="7ed001b4236e4af6e20b2800c11037beeadd23bc12a419e3769d0ca9cc7e4d2a" exitCode=0 Sep 30 16:40:59 crc kubenswrapper[4796]: I0930 16:40:59.945762 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" event={"ID":"743c014a-a242-4a7a-96e5-167a0d0928fb","Type":"ContainerDied","Data":"7ed001b4236e4af6e20b2800c11037beeadd23bc12a419e3769d0ca9cc7e4d2a"} Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.418241 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.503790 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-ssh-key\") pod \"743c014a-a242-4a7a-96e5-167a0d0928fb\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.503906 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-inventory\") pod \"743c014a-a242-4a7a-96e5-167a0d0928fb\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.541459 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "743c014a-a242-4a7a-96e5-167a0d0928fb" (UID: "743c014a-a242-4a7a-96e5-167a0d0928fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.551492 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-inventory" (OuterVolumeSpecName: "inventory") pod "743c014a-a242-4a7a-96e5-167a0d0928fb" (UID: "743c014a-a242-4a7a-96e5-167a0d0928fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.605276 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4jtc\" (UniqueName: \"kubernetes.io/projected/743c014a-a242-4a7a-96e5-167a0d0928fb-kube-api-access-d4jtc\") pod \"743c014a-a242-4a7a-96e5-167a0d0928fb\" (UID: \"743c014a-a242-4a7a-96e5-167a0d0928fb\") " Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.605897 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.605928 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/743c014a-a242-4a7a-96e5-167a0d0928fb-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.609849 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743c014a-a242-4a7a-96e5-167a0d0928fb-kube-api-access-d4jtc" (OuterVolumeSpecName: "kube-api-access-d4jtc") pod "743c014a-a242-4a7a-96e5-167a0d0928fb" (UID: "743c014a-a242-4a7a-96e5-167a0d0928fb"). InnerVolumeSpecName "kube-api-access-d4jtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.707381 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4jtc\" (UniqueName: \"kubernetes.io/projected/743c014a-a242-4a7a-96e5-167a0d0928fb-kube-api-access-d4jtc\") on node \"crc\" DevicePath \"\"" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.969585 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" event={"ID":"743c014a-a242-4a7a-96e5-167a0d0928fb","Type":"ContainerDied","Data":"c518371290c2cdf582d72d1d1a74e3a2c468dc57d2511dd46f1ceec926497c2d"} Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.969629 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c518371290c2cdf582d72d1d1a74e3a2c468dc57d2511dd46f1ceec926497c2d" Sep 30 16:41:01 crc kubenswrapper[4796]: I0930 16:41:01.969674 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zm5lv" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063031 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx"] Sep 30 16:41:02 crc kubenswrapper[4796]: E0930 16:41:02.063494 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743c014a-a242-4a7a-96e5-167a0d0928fb" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063517 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="743c014a-a242-4a7a-96e5-167a0d0928fb" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:41:02 crc kubenswrapper[4796]: E0930 16:41:02.063533 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="registry-server" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063541 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="registry-server" Sep 30 16:41:02 crc kubenswrapper[4796]: E0930 16:41:02.063567 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="extract-utilities" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063574 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="extract-utilities" Sep 30 16:41:02 crc kubenswrapper[4796]: E0930 16:41:02.063595 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="extract-content" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063601 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="extract-content" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063833 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="970d4ab4-bebb-4fb0-b000-854bd7ed9e99" containerName="registry-server" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.063854 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="743c014a-a242-4a7a-96e5-167a0d0928fb" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.064685 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.070928 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.071139 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.071300 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.071341 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.076830 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx"] Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.216674 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.216851 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brl55\" (UniqueName: \"kubernetes.io/projected/2c6044b7-9944-42ac-821a-8ddad26ad823-kube-api-access-brl55\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.216885 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.318467 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brl55\" (UniqueName: \"kubernetes.io/projected/2c6044b7-9944-42ac-821a-8ddad26ad823-kube-api-access-brl55\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.318788 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.318878 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.322780 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.330690 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.336858 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brl55\" (UniqueName: \"kubernetes.io/projected/2c6044b7-9944-42ac-821a-8ddad26ad823-kube-api-access-brl55\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-r94sx\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.383007 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.910952 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx"] Sep 30 16:41:02 crc kubenswrapper[4796]: I0930 16:41:02.980672 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" event={"ID":"2c6044b7-9944-42ac-821a-8ddad26ad823","Type":"ContainerStarted","Data":"e06ba789111232c3c3745dc6311e02ffd980b0ef0c2c3ea5ebf855a2a8b0ea8f"} Sep 30 16:41:03 crc kubenswrapper[4796]: I0930 16:41:03.995632 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" event={"ID":"2c6044b7-9944-42ac-821a-8ddad26ad823","Type":"ContainerStarted","Data":"fa18bb2f6b9f10cb8afac3baed87bc1bd0c2b4c14acd6a947bd07752a11ac063"} Sep 30 16:41:04 crc kubenswrapper[4796]: I0930 16:41:04.028427 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" podStartSLOduration=1.546641025 podStartE2EDuration="2.028405172s" podCreationTimestamp="2025-09-30 16:41:02 +0000 UTC" firstStartedPulling="2025-09-30 16:41:02.911998644 +0000 UTC m=+1754.925277171" lastFinishedPulling="2025-09-30 16:41:03.393762781 +0000 UTC m=+1755.407041318" observedRunningTime="2025-09-30 16:41:04.015583943 +0000 UTC m=+1756.028862500" watchObservedRunningTime="2025-09-30 16:41:04.028405172 +0000 UTC m=+1756.041683719" Sep 30 16:41:10 crc kubenswrapper[4796]: I0930 16:41:10.733274 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:41:10 crc kubenswrapper[4796]: E0930 16:41:10.734164 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:41:12 crc kubenswrapper[4796]: I0930 16:41:12.075018 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-q8zkj"] Sep 30 16:41:12 crc kubenswrapper[4796]: I0930 16:41:12.088044 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-q8zkj"] Sep 30 16:41:12 crc kubenswrapper[4796]: I0930 16:41:12.747339 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4" path="/var/lib/kubelet/pods/c6da3535-7e08-4b79-a4d5-6c94ab7ff0a4/volumes" Sep 30 16:41:24 crc kubenswrapper[4796]: I0930 16:41:24.733657 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:41:24 crc kubenswrapper[4796]: E0930 16:41:24.734759 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:41:39 crc kubenswrapper[4796]: I0930 16:41:39.733205 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:41:40 crc kubenswrapper[4796]: I0930 16:41:40.373719 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c"} Sep 30 16:41:55 crc kubenswrapper[4796]: I0930 16:41:55.186415 4796 scope.go:117] "RemoveContainer" containerID="3f039a86e401429878c93499cf246807989661b44eded65d71b2bfc3b7e90787" Sep 30 16:41:55 crc kubenswrapper[4796]: I0930 16:41:55.222108 4796 scope.go:117] "RemoveContainer" containerID="ac54fdb717ec98f5bc5aa9bede4e1e7f76dd72191db4ea4639250c35c4d24508" Sep 30 16:41:55 crc kubenswrapper[4796]: I0930 16:41:55.308509 4796 scope.go:117] "RemoveContainer" containerID="7dd05f77c0a5e66b6021d4cce406dda197589a232df375e5cea5aa8f59debc68" Sep 30 16:41:55 crc kubenswrapper[4796]: I0930 16:41:55.341754 4796 scope.go:117] "RemoveContainer" containerID="71aca77e8fb729d6597f8a37c743a6b1880369f53caf140fdcba5a93bea7d9d1" Sep 30 16:42:01 crc kubenswrapper[4796]: I0930 16:42:01.584175 4796 generic.go:334] "Generic (PLEG): container finished" podID="2c6044b7-9944-42ac-821a-8ddad26ad823" containerID="fa18bb2f6b9f10cb8afac3baed87bc1bd0c2b4c14acd6a947bd07752a11ac063" exitCode=2 Sep 30 16:42:01 crc kubenswrapper[4796]: I0930 16:42:01.584255 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" event={"ID":"2c6044b7-9944-42ac-821a-8ddad26ad823","Type":"ContainerDied","Data":"fa18bb2f6b9f10cb8afac3baed87bc1bd0c2b4c14acd6a947bd07752a11ac063"} Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.041497 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.134837 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-inventory\") pod \"2c6044b7-9944-42ac-821a-8ddad26ad823\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.135007 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brl55\" (UniqueName: \"kubernetes.io/projected/2c6044b7-9944-42ac-821a-8ddad26ad823-kube-api-access-brl55\") pod \"2c6044b7-9944-42ac-821a-8ddad26ad823\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.135158 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-ssh-key\") pod \"2c6044b7-9944-42ac-821a-8ddad26ad823\" (UID: \"2c6044b7-9944-42ac-821a-8ddad26ad823\") " Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.143883 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c6044b7-9944-42ac-821a-8ddad26ad823-kube-api-access-brl55" (OuterVolumeSpecName: "kube-api-access-brl55") pod "2c6044b7-9944-42ac-821a-8ddad26ad823" (UID: "2c6044b7-9944-42ac-821a-8ddad26ad823"). InnerVolumeSpecName "kube-api-access-brl55". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.167133 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c6044b7-9944-42ac-821a-8ddad26ad823" (UID: "2c6044b7-9944-42ac-821a-8ddad26ad823"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.185504 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-inventory" (OuterVolumeSpecName: "inventory") pod "2c6044b7-9944-42ac-821a-8ddad26ad823" (UID: "2c6044b7-9944-42ac-821a-8ddad26ad823"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.237318 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brl55\" (UniqueName: \"kubernetes.io/projected/2c6044b7-9944-42ac-821a-8ddad26ad823-kube-api-access-brl55\") on node \"crc\" DevicePath \"\"" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.237348 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.237361 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c6044b7-9944-42ac-821a-8ddad26ad823-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.605710 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" event={"ID":"2c6044b7-9944-42ac-821a-8ddad26ad823","Type":"ContainerDied","Data":"e06ba789111232c3c3745dc6311e02ffd980b0ef0c2c3ea5ebf855a2a8b0ea8f"} Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.605756 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-r94sx" Sep 30 16:42:03 crc kubenswrapper[4796]: I0930 16:42:03.605760 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e06ba789111232c3c3745dc6311e02ffd980b0ef0c2c3ea5ebf855a2a8b0ea8f" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.042967 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7"] Sep 30 16:42:11 crc kubenswrapper[4796]: E0930 16:42:11.044142 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6044b7-9944-42ac-821a-8ddad26ad823" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.044161 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6044b7-9944-42ac-821a-8ddad26ad823" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.044404 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c6044b7-9944-42ac-821a-8ddad26ad823" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.045207 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.052830 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.053175 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.053175 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.053190 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.065755 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7"] Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.190966 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-kube-api-access-r5q6t\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.191158 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.191281 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.293463 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-kube-api-access-r5q6t\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.293555 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.293615 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.300134 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.306577 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.313955 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-kube-api-access-r5q6t\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.368087 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.916544 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7"] Sep 30 16:42:11 crc kubenswrapper[4796]: I0930 16:42:11.927354 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:42:12 crc kubenswrapper[4796]: I0930 16:42:12.690550 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" event={"ID":"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2","Type":"ContainerStarted","Data":"004cdc21d32071748783150e5f4d1ebb4eedd4a745b72416e68ce3fda8610985"} Sep 30 16:42:12 crc kubenswrapper[4796]: I0930 16:42:12.690786 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" event={"ID":"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2","Type":"ContainerStarted","Data":"b77df2186893e8846a674d45b5c919a3664efad95330299584c89bff744563ee"} Sep 30 16:42:12 crc kubenswrapper[4796]: I0930 16:42:12.728335 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" podStartSLOduration=1.204910118 podStartE2EDuration="1.728315265s" podCreationTimestamp="2025-09-30 16:42:11 +0000 UTC" firstStartedPulling="2025-09-30 16:42:11.927072465 +0000 UTC m=+1823.940350992" lastFinishedPulling="2025-09-30 16:42:12.450477602 +0000 UTC m=+1824.463756139" observedRunningTime="2025-09-30 16:42:12.710355628 +0000 UTC m=+1824.723634155" watchObservedRunningTime="2025-09-30 16:42:12.728315265 +0000 UTC m=+1824.741593802" Sep 30 16:43:02 crc kubenswrapper[4796]: I0930 16:43:02.207560 4796 generic.go:334] "Generic (PLEG): container finished" podID="27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" containerID="004cdc21d32071748783150e5f4d1ebb4eedd4a745b72416e68ce3fda8610985" exitCode=0 Sep 30 16:43:02 crc kubenswrapper[4796]: I0930 16:43:02.208178 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" event={"ID":"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2","Type":"ContainerDied","Data":"004cdc21d32071748783150e5f4d1ebb4eedd4a745b72416e68ce3fda8610985"} Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.695108 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.733710 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-kube-api-access-r5q6t\") pod \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.733809 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-ssh-key\") pod \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.733850 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-inventory\") pod \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\" (UID: \"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2\") " Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.748060 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-kube-api-access-r5q6t" (OuterVolumeSpecName: "kube-api-access-r5q6t") pod "27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" (UID: "27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2"). InnerVolumeSpecName "kube-api-access-r5q6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.764539 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" (UID: "27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.764855 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-inventory" (OuterVolumeSpecName: "inventory") pod "27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" (UID: "27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.835656 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.835687 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:03 crc kubenswrapper[4796]: I0930 16:43:03.835699 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2-kube-api-access-r5q6t\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.235942 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" event={"ID":"27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2","Type":"ContainerDied","Data":"b77df2186893e8846a674d45b5c919a3664efad95330299584c89bff744563ee"} Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.236021 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b77df2186893e8846a674d45b5c919a3664efad95330299584c89bff744563ee" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.236115 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.349530 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-j5xtv"] Sep 30 16:43:04 crc kubenswrapper[4796]: E0930 16:43:04.350268 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.350301 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.350680 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.351838 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.356562 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.356940 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.358139 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.358401 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.383766 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-j5xtv"] Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.550921 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.551057 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.551294 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfdrm\" (UniqueName: \"kubernetes.io/projected/35c5d020-3189-41c5-90f4-4544c251539e-kube-api-access-wfdrm\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.654652 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.654784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.654881 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfdrm\" (UniqueName: \"kubernetes.io/projected/35c5d020-3189-41c5-90f4-4544c251539e-kube-api-access-wfdrm\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.659755 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.661348 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.683194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfdrm\" (UniqueName: \"kubernetes.io/projected/35c5d020-3189-41c5-90f4-4544c251539e-kube-api-access-wfdrm\") pod \"ssh-known-hosts-edpm-deployment-j5xtv\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:04 crc kubenswrapper[4796]: I0930 16:43:04.685884 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:05 crc kubenswrapper[4796]: I0930 16:43:05.274595 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-j5xtv"] Sep 30 16:43:06 crc kubenswrapper[4796]: I0930 16:43:06.254737 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" event={"ID":"35c5d020-3189-41c5-90f4-4544c251539e","Type":"ContainerStarted","Data":"b5ca31ad5fce3c145e7d4e924d58ff863c5fb6b8f8491c370e9ccc17db4d8cb6"} Sep 30 16:43:06 crc kubenswrapper[4796]: I0930 16:43:06.255219 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" event={"ID":"35c5d020-3189-41c5-90f4-4544c251539e","Type":"ContainerStarted","Data":"f0547c70928f84c0911b342dfbfdf6feb876c89162ca31a2f295ac8064d32515"} Sep 30 16:43:06 crc kubenswrapper[4796]: I0930 16:43:06.280291 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" podStartSLOduration=1.8288012949999999 podStartE2EDuration="2.280254908s" podCreationTimestamp="2025-09-30 16:43:04 +0000 UTC" firstStartedPulling="2025-09-30 16:43:05.281384956 +0000 UTC m=+1877.294663503" lastFinishedPulling="2025-09-30 16:43:05.732838549 +0000 UTC m=+1877.746117116" observedRunningTime="2025-09-30 16:43:06.276895551 +0000 UTC m=+1878.290174108" watchObservedRunningTime="2025-09-30 16:43:06.280254908 +0000 UTC m=+1878.293533435" Sep 30 16:43:13 crc kubenswrapper[4796]: I0930 16:43:13.314542 4796 generic.go:334] "Generic (PLEG): container finished" podID="35c5d020-3189-41c5-90f4-4544c251539e" containerID="b5ca31ad5fce3c145e7d4e924d58ff863c5fb6b8f8491c370e9ccc17db4d8cb6" exitCode=0 Sep 30 16:43:13 crc kubenswrapper[4796]: I0930 16:43:13.314815 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" event={"ID":"35c5d020-3189-41c5-90f4-4544c251539e","Type":"ContainerDied","Data":"b5ca31ad5fce3c145e7d4e924d58ff863c5fb6b8f8491c370e9ccc17db4d8cb6"} Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.775135 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.952187 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-inventory-0\") pod \"35c5d020-3189-41c5-90f4-4544c251539e\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.952351 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfdrm\" (UniqueName: \"kubernetes.io/projected/35c5d020-3189-41c5-90f4-4544c251539e-kube-api-access-wfdrm\") pod \"35c5d020-3189-41c5-90f4-4544c251539e\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.952460 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-ssh-key-openstack-edpm-ipam\") pod \"35c5d020-3189-41c5-90f4-4544c251539e\" (UID: \"35c5d020-3189-41c5-90f4-4544c251539e\") " Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.959656 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c5d020-3189-41c5-90f4-4544c251539e-kube-api-access-wfdrm" (OuterVolumeSpecName: "kube-api-access-wfdrm") pod "35c5d020-3189-41c5-90f4-4544c251539e" (UID: "35c5d020-3189-41c5-90f4-4544c251539e"). InnerVolumeSpecName "kube-api-access-wfdrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.988630 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "35c5d020-3189-41c5-90f4-4544c251539e" (UID: "35c5d020-3189-41c5-90f4-4544c251539e"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:14 crc kubenswrapper[4796]: I0930 16:43:14.994472 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "35c5d020-3189-41c5-90f4-4544c251539e" (UID: "35c5d020-3189-41c5-90f4-4544c251539e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.054904 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfdrm\" (UniqueName: \"kubernetes.io/projected/35c5d020-3189-41c5-90f4-4544c251539e-kube-api-access-wfdrm\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.054951 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.054965 4796 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35c5d020-3189-41c5-90f4-4544c251539e-inventory-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.334688 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" event={"ID":"35c5d020-3189-41c5-90f4-4544c251539e","Type":"ContainerDied","Data":"f0547c70928f84c0911b342dfbfdf6feb876c89162ca31a2f295ac8064d32515"} Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.334734 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0547c70928f84c0911b342dfbfdf6feb876c89162ca31a2f295ac8064d32515" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.334762 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j5xtv" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.409381 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x"] Sep 30 16:43:15 crc kubenswrapper[4796]: E0930 16:43:15.410973 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c5d020-3189-41c5-90f4-4544c251539e" containerName="ssh-known-hosts-edpm-deployment" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.411065 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c5d020-3189-41c5-90f4-4544c251539e" containerName="ssh-known-hosts-edpm-deployment" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.411238 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c5d020-3189-41c5-90f4-4544c251539e" containerName="ssh-known-hosts-edpm-deployment" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.412076 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.415356 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.415581 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.415869 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.426125 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.433819 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x"] Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.565876 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltrk8\" (UniqueName: \"kubernetes.io/projected/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-kube-api-access-ltrk8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.566164 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.566267 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.668090 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.668291 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.670607 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltrk8\" (UniqueName: \"kubernetes.io/projected/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-kube-api-access-ltrk8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.672406 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.673678 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.690382 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltrk8\" (UniqueName: \"kubernetes.io/projected/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-kube-api-access-ltrk8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q2j8x\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:15 crc kubenswrapper[4796]: I0930 16:43:15.731819 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:16 crc kubenswrapper[4796]: I0930 16:43:16.294663 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x"] Sep 30 16:43:16 crc kubenswrapper[4796]: W0930 16:43:16.301117 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1351ba7c_3e3e_42ff_960f_8ada3d7e490a.slice/crio-adf6fa026d8ddf95921038c74e73297dfc8bf275ad672fdc521d0b4883bf41d8 WatchSource:0}: Error finding container adf6fa026d8ddf95921038c74e73297dfc8bf275ad672fdc521d0b4883bf41d8: Status 404 returned error can't find the container with id adf6fa026d8ddf95921038c74e73297dfc8bf275ad672fdc521d0b4883bf41d8 Sep 30 16:43:16 crc kubenswrapper[4796]: I0930 16:43:16.347209 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" event={"ID":"1351ba7c-3e3e-42ff-960f-8ada3d7e490a","Type":"ContainerStarted","Data":"adf6fa026d8ddf95921038c74e73297dfc8bf275ad672fdc521d0b4883bf41d8"} Sep 30 16:43:17 crc kubenswrapper[4796]: I0930 16:43:17.359821 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" event={"ID":"1351ba7c-3e3e-42ff-960f-8ada3d7e490a","Type":"ContainerStarted","Data":"6e731a1da13626822979eccf68f12d18c041db9763a0c4ed424b21270c379d9f"} Sep 30 16:43:17 crc kubenswrapper[4796]: I0930 16:43:17.383636 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" podStartSLOduration=1.780510888 podStartE2EDuration="2.383610151s" podCreationTimestamp="2025-09-30 16:43:15 +0000 UTC" firstStartedPulling="2025-09-30 16:43:16.304559028 +0000 UTC m=+1888.317837555" lastFinishedPulling="2025-09-30 16:43:16.907658261 +0000 UTC m=+1888.920936818" observedRunningTime="2025-09-30 16:43:17.379009288 +0000 UTC m=+1889.392287855" watchObservedRunningTime="2025-09-30 16:43:17.383610151 +0000 UTC m=+1889.396888718" Sep 30 16:43:26 crc kubenswrapper[4796]: I0930 16:43:26.445430 4796 generic.go:334] "Generic (PLEG): container finished" podID="1351ba7c-3e3e-42ff-960f-8ada3d7e490a" containerID="6e731a1da13626822979eccf68f12d18c041db9763a0c4ed424b21270c379d9f" exitCode=0 Sep 30 16:43:26 crc kubenswrapper[4796]: I0930 16:43:26.445476 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" event={"ID":"1351ba7c-3e3e-42ff-960f-8ada3d7e490a","Type":"ContainerDied","Data":"6e731a1da13626822979eccf68f12d18c041db9763a0c4ed424b21270c379d9f"} Sep 30 16:43:27 crc kubenswrapper[4796]: I0930 16:43:27.852322 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.004793 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltrk8\" (UniqueName: \"kubernetes.io/projected/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-kube-api-access-ltrk8\") pod \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.005059 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-ssh-key\") pod \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.005153 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-inventory\") pod \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\" (UID: \"1351ba7c-3e3e-42ff-960f-8ada3d7e490a\") " Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.010067 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-kube-api-access-ltrk8" (OuterVolumeSpecName: "kube-api-access-ltrk8") pod "1351ba7c-3e3e-42ff-960f-8ada3d7e490a" (UID: "1351ba7c-3e3e-42ff-960f-8ada3d7e490a"). InnerVolumeSpecName "kube-api-access-ltrk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.031275 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-inventory" (OuterVolumeSpecName: "inventory") pod "1351ba7c-3e3e-42ff-960f-8ada3d7e490a" (UID: "1351ba7c-3e3e-42ff-960f-8ada3d7e490a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.040358 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1351ba7c-3e3e-42ff-960f-8ada3d7e490a" (UID: "1351ba7c-3e3e-42ff-960f-8ada3d7e490a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.107610 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.107672 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.107719 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltrk8\" (UniqueName: \"kubernetes.io/projected/1351ba7c-3e3e-42ff-960f-8ada3d7e490a-kube-api-access-ltrk8\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.466668 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" event={"ID":"1351ba7c-3e3e-42ff-960f-8ada3d7e490a","Type":"ContainerDied","Data":"adf6fa026d8ddf95921038c74e73297dfc8bf275ad672fdc521d0b4883bf41d8"} Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.466924 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adf6fa026d8ddf95921038c74e73297dfc8bf275ad672fdc521d0b4883bf41d8" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.466756 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q2j8x" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.531750 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2"] Sep 30 16:43:28 crc kubenswrapper[4796]: E0930 16:43:28.532207 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1351ba7c-3e3e-42ff-960f-8ada3d7e490a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.532224 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1351ba7c-3e3e-42ff-960f-8ada3d7e490a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.532413 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1351ba7c-3e3e-42ff-960f-8ada3d7e490a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.533015 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.535527 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.535531 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.535952 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.536148 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.557493 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2"] Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.617332 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.617570 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.617744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d8sf\" (UniqueName: \"kubernetes.io/projected/8b67950f-98c0-44fe-a09a-77e356e069ff-kube-api-access-2d8sf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.719090 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.719259 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.719374 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d8sf\" (UniqueName: \"kubernetes.io/projected/8b67950f-98c0-44fe-a09a-77e356e069ff-kube-api-access-2d8sf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.724618 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.725377 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.738381 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d8sf\" (UniqueName: \"kubernetes.io/projected/8b67950f-98c0-44fe-a09a-77e356e069ff-kube-api-access-2d8sf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:28 crc kubenswrapper[4796]: I0930 16:43:28.869221 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:29 crc kubenswrapper[4796]: I0930 16:43:29.407197 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2"] Sep 30 16:43:29 crc kubenswrapper[4796]: I0930 16:43:29.477111 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" event={"ID":"8b67950f-98c0-44fe-a09a-77e356e069ff","Type":"ContainerStarted","Data":"5a0a0dbd9d9ddb250c8f522e5e57e6139a3e8851b48216c9af5f397258fccc05"} Sep 30 16:43:30 crc kubenswrapper[4796]: I0930 16:43:30.487178 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" event={"ID":"8b67950f-98c0-44fe-a09a-77e356e069ff","Type":"ContainerStarted","Data":"fbf2536ddb243c3563bfcf440d34625cf1e6372dab6a3cd4b69880c0d7711fde"} Sep 30 16:43:30 crc kubenswrapper[4796]: I0930 16:43:30.509307 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" podStartSLOduration=1.7336584739999998 podStartE2EDuration="2.509272766s" podCreationTimestamp="2025-09-30 16:43:28 +0000 UTC" firstStartedPulling="2025-09-30 16:43:29.408309702 +0000 UTC m=+1901.421588229" lastFinishedPulling="2025-09-30 16:43:30.183923984 +0000 UTC m=+1902.197202521" observedRunningTime="2025-09-30 16:43:30.500442241 +0000 UTC m=+1902.513720768" watchObservedRunningTime="2025-09-30 16:43:30.509272766 +0000 UTC m=+1902.522551283" Sep 30 16:43:40 crc kubenswrapper[4796]: I0930 16:43:40.586380 4796 generic.go:334] "Generic (PLEG): container finished" podID="8b67950f-98c0-44fe-a09a-77e356e069ff" containerID="fbf2536ddb243c3563bfcf440d34625cf1e6372dab6a3cd4b69880c0d7711fde" exitCode=0 Sep 30 16:43:40 crc kubenswrapper[4796]: I0930 16:43:40.586433 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" event={"ID":"8b67950f-98c0-44fe-a09a-77e356e069ff","Type":"ContainerDied","Data":"fbf2536ddb243c3563bfcf440d34625cf1e6372dab6a3cd4b69880c0d7711fde"} Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.099180 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.294837 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d8sf\" (UniqueName: \"kubernetes.io/projected/8b67950f-98c0-44fe-a09a-77e356e069ff-kube-api-access-2d8sf\") pod \"8b67950f-98c0-44fe-a09a-77e356e069ff\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.295452 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-ssh-key\") pod \"8b67950f-98c0-44fe-a09a-77e356e069ff\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.296048 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-inventory\") pod \"8b67950f-98c0-44fe-a09a-77e356e069ff\" (UID: \"8b67950f-98c0-44fe-a09a-77e356e069ff\") " Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.301902 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b67950f-98c0-44fe-a09a-77e356e069ff-kube-api-access-2d8sf" (OuterVolumeSpecName: "kube-api-access-2d8sf") pod "8b67950f-98c0-44fe-a09a-77e356e069ff" (UID: "8b67950f-98c0-44fe-a09a-77e356e069ff"). InnerVolumeSpecName "kube-api-access-2d8sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.322584 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-inventory" (OuterVolumeSpecName: "inventory") pod "8b67950f-98c0-44fe-a09a-77e356e069ff" (UID: "8b67950f-98c0-44fe-a09a-77e356e069ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.325642 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b67950f-98c0-44fe-a09a-77e356e069ff" (UID: "8b67950f-98c0-44fe-a09a-77e356e069ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.398664 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.398720 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b67950f-98c0-44fe-a09a-77e356e069ff-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.398739 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d8sf\" (UniqueName: \"kubernetes.io/projected/8b67950f-98c0-44fe-a09a-77e356e069ff-kube-api-access-2d8sf\") on node \"crc\" DevicePath \"\"" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.611134 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" event={"ID":"8b67950f-98c0-44fe-a09a-77e356e069ff","Type":"ContainerDied","Data":"5a0a0dbd9d9ddb250c8f522e5e57e6139a3e8851b48216c9af5f397258fccc05"} Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.611187 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a0a0dbd9d9ddb250c8f522e5e57e6139a3e8851b48216c9af5f397258fccc05" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.611276 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.700135 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc"] Sep 30 16:43:42 crc kubenswrapper[4796]: E0930 16:43:42.700503 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b67950f-98c0-44fe-a09a-77e356e069ff" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.700519 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b67950f-98c0-44fe-a09a-77e356e069ff" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.700772 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b67950f-98c0-44fe-a09a-77e356e069ff" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.701456 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704128 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704437 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704532 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mzd8\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-kube-api-access-5mzd8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704606 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704643 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704696 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704771 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704817 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704844 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.704953 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.705027 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.705075 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.705102 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.705162 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.705290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.705462 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.707671 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.709149 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.710105 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.711091 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.711290 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.713258 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.720662 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc"] Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.806616 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.806680 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.806712 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mzd8\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-kube-api-access-5mzd8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.806807 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.806839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.806925 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807012 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807047 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807070 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807103 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807144 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807176 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807200 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.807248 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.811563 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.812115 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.812942 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.813009 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.814515 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.814800 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.815556 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.815660 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.816118 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.816828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.817163 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.817792 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.822331 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:42 crc kubenswrapper[4796]: I0930 16:43:42.830171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mzd8\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-kube-api-access-5mzd8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:43 crc kubenswrapper[4796]: I0930 16:43:43.022966 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:43:43 crc kubenswrapper[4796]: I0930 16:43:43.562803 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc"] Sep 30 16:43:43 crc kubenswrapper[4796]: I0930 16:43:43.631222 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" event={"ID":"185ba20f-66a9-4425-a1ba-bfc42a8f1c01","Type":"ContainerStarted","Data":"245ad5cf79419e81952cb4b53a7be1cea66535f8c79a71c62ff5ddc5357d4e27"} Sep 30 16:43:44 crc kubenswrapper[4796]: I0930 16:43:44.643616 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" event={"ID":"185ba20f-66a9-4425-a1ba-bfc42a8f1c01","Type":"ContainerStarted","Data":"709f94afb4d46fa8f4c9a0666038f0929f72c8e6e08bf02c62d7fe8077fd1f24"} Sep 30 16:43:44 crc kubenswrapper[4796]: I0930 16:43:44.671578 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" podStartSLOduration=2.184763615 podStartE2EDuration="2.671559398s" podCreationTimestamp="2025-09-30 16:43:42 +0000 UTC" firstStartedPulling="2025-09-30 16:43:43.572251912 +0000 UTC m=+1915.585530449" lastFinishedPulling="2025-09-30 16:43:44.059047695 +0000 UTC m=+1916.072326232" observedRunningTime="2025-09-30 16:43:44.666718919 +0000 UTC m=+1916.679997466" watchObservedRunningTime="2025-09-30 16:43:44.671559398 +0000 UTC m=+1916.684837935" Sep 30 16:44:01 crc kubenswrapper[4796]: I0930 16:44:01.097294 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:44:01 crc kubenswrapper[4796]: I0930 16:44:01.098152 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:44:27 crc kubenswrapper[4796]: I0930 16:44:27.077367 4796 generic.go:334] "Generic (PLEG): container finished" podID="185ba20f-66a9-4425-a1ba-bfc42a8f1c01" containerID="709f94afb4d46fa8f4c9a0666038f0929f72c8e6e08bf02c62d7fe8077fd1f24" exitCode=0 Sep 30 16:44:27 crc kubenswrapper[4796]: I0930 16:44:27.077502 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" event={"ID":"185ba20f-66a9-4425-a1ba-bfc42a8f1c01","Type":"ContainerDied","Data":"709f94afb4d46fa8f4c9a0666038f0929f72c8e6e08bf02c62d7fe8077fd1f24"} Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.504053 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654453 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ssh-key\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654529 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654568 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-inventory\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654603 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-nova-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654623 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-repo-setup-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654680 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-libvirt-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654722 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-bootstrap-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654774 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ovn-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654799 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-ovn-default-certs-0\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654864 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-telemetry-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654898 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.654920 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.655040 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mzd8\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-kube-api-access-5mzd8\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.655092 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-neutron-metadata-combined-ca-bundle\") pod \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\" (UID: \"185ba20f-66a9-4425-a1ba-bfc42a8f1c01\") " Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.661533 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.662453 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.662853 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.663577 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.664138 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.665803 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.666064 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-kube-api-access-5mzd8" (OuterVolumeSpecName: "kube-api-access-5mzd8") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "kube-api-access-5mzd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.666163 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.666245 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.667718 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.667781 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.676312 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.690403 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.697172 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-inventory" (OuterVolumeSpecName: "inventory") pod "185ba20f-66a9-4425-a1ba-bfc42a8f1c01" (UID: "185ba20f-66a9-4425-a1ba-bfc42a8f1c01"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.760159 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.760227 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761009 4796 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761186 4796 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761222 4796 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761251 4796 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761293 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761322 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761384 4796 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761414 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761455 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761475 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mzd8\" (UniqueName: \"kubernetes.io/projected/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-kube-api-access-5mzd8\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761494 4796 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:28 crc kubenswrapper[4796]: I0930 16:44:28.761515 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/185ba20f-66a9-4425-a1ba-bfc42a8f1c01-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.097246 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" event={"ID":"185ba20f-66a9-4425-a1ba-bfc42a8f1c01","Type":"ContainerDied","Data":"245ad5cf79419e81952cb4b53a7be1cea66535f8c79a71c62ff5ddc5357d4e27"} Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.097286 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="245ad5cf79419e81952cb4b53a7be1cea66535f8c79a71c62ff5ddc5357d4e27" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.097335 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.259081 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88"] Sep 30 16:44:29 crc kubenswrapper[4796]: E0930 16:44:29.259441 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185ba20f-66a9-4425-a1ba-bfc42a8f1c01" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.259459 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="185ba20f-66a9-4425-a1ba-bfc42a8f1c01" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.259623 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="185ba20f-66a9-4425-a1ba-bfc42a8f1c01" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.260228 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.261973 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.262151 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.262517 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.263269 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.264160 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.273282 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88"] Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.376058 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.376158 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.376231 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg8xs\" (UniqueName: \"kubernetes.io/projected/abb53121-c532-478a-9670-8350dd3e477b-kube-api-access-bg8xs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.376319 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.376601 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/abb53121-c532-478a-9670-8350dd3e477b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.478199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/abb53121-c532-478a-9670-8350dd3e477b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.478282 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.478330 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.478373 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg8xs\" (UniqueName: \"kubernetes.io/projected/abb53121-c532-478a-9670-8350dd3e477b-kube-api-access-bg8xs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.478435 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.479153 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/abb53121-c532-478a-9670-8350dd3e477b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.482420 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.482649 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.487106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.502226 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg8xs\" (UniqueName: \"kubernetes.io/projected/abb53121-c532-478a-9670-8350dd3e477b-kube-api-access-bg8xs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vtt88\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:29 crc kubenswrapper[4796]: I0930 16:44:29.579727 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:44:30 crc kubenswrapper[4796]: I0930 16:44:30.157965 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88"] Sep 30 16:44:31 crc kubenswrapper[4796]: I0930 16:44:31.097338 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:44:31 crc kubenswrapper[4796]: I0930 16:44:31.097905 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:44:31 crc kubenswrapper[4796]: I0930 16:44:31.113402 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" event={"ID":"abb53121-c532-478a-9670-8350dd3e477b","Type":"ContainerStarted","Data":"a210593a4a3c83f342c2bb7a928dd68e60e92c1d5b38f01a8dcfd1b92baced06"} Sep 30 16:44:32 crc kubenswrapper[4796]: I0930 16:44:32.123507 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" event={"ID":"abb53121-c532-478a-9670-8350dd3e477b","Type":"ContainerStarted","Data":"6403599bedb94f41bf83ded6db807312a66aa633747940ba3173fe5f69db1673"} Sep 30 16:44:32 crc kubenswrapper[4796]: I0930 16:44:32.169552 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" podStartSLOduration=2.207751055 podStartE2EDuration="3.169531057s" podCreationTimestamp="2025-09-30 16:44:29 +0000 UTC" firstStartedPulling="2025-09-30 16:44:30.165215374 +0000 UTC m=+1962.178493901" lastFinishedPulling="2025-09-30 16:44:31.126995386 +0000 UTC m=+1963.140273903" observedRunningTime="2025-09-30 16:44:32.156965675 +0000 UTC m=+1964.170244222" watchObservedRunningTime="2025-09-30 16:44:32.169531057 +0000 UTC m=+1964.182809584" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.147639 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6"] Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.150211 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.152861 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.153929 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.176479 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6"] Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.306167 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjvmm\" (UniqueName: \"kubernetes.io/projected/6352d300-c932-4544-918f-4ccb6b5bb2eb-kube-api-access-sjvmm\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.306268 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6352d300-c932-4544-918f-4ccb6b5bb2eb-secret-volume\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.306304 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6352d300-c932-4544-918f-4ccb6b5bb2eb-config-volume\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.407518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6352d300-c932-4544-918f-4ccb6b5bb2eb-secret-volume\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.407579 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6352d300-c932-4544-918f-4ccb6b5bb2eb-config-volume\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.407711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjvmm\" (UniqueName: \"kubernetes.io/projected/6352d300-c932-4544-918f-4ccb6b5bb2eb-kube-api-access-sjvmm\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.408752 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6352d300-c932-4544-918f-4ccb6b5bb2eb-config-volume\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.425339 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjvmm\" (UniqueName: \"kubernetes.io/projected/6352d300-c932-4544-918f-4ccb6b5bb2eb-kube-api-access-sjvmm\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.429194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6352d300-c932-4544-918f-4ccb6b5bb2eb-secret-volume\") pod \"collect-profiles-29320845-qnpf6\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.475041 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:00 crc kubenswrapper[4796]: I0930 16:45:00.911612 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6"] Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.096593 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.096831 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.097036 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.098013 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.098196 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c" gracePeriod=600 Sep 30 16:45:01 crc kubenswrapper[4796]: E0930 16:45:01.260485 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod670c655e_3953_4773_84ef_19c678d482f9.slice/crio-19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.442599 4796 generic.go:334] "Generic (PLEG): container finished" podID="6352d300-c932-4544-918f-4ccb6b5bb2eb" containerID="04d2bea011c92236d9d4cf7cd7ff54099f2756548b30cd2baa0421e0b97afa82" exitCode=0 Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.442650 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" event={"ID":"6352d300-c932-4544-918f-4ccb6b5bb2eb","Type":"ContainerDied","Data":"04d2bea011c92236d9d4cf7cd7ff54099f2756548b30cd2baa0421e0b97afa82"} Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.442713 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" event={"ID":"6352d300-c932-4544-918f-4ccb6b5bb2eb","Type":"ContainerStarted","Data":"e17a51e768534c0cd1d8301bc9ba35b55da587e6be3b9500490bc4682f37c687"} Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.445766 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c" exitCode=0 Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.445817 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c"} Sep 30 16:45:01 crc kubenswrapper[4796]: I0930 16:45:01.445856 4796 scope.go:117] "RemoveContainer" containerID="ba8eb37eaf83995785e664aeb671a80aa02fec9429b7afb7c6acc5cc54683fd2" Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.459862 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752"} Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.801447 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.959758 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6352d300-c932-4544-918f-4ccb6b5bb2eb-config-volume\") pod \"6352d300-c932-4544-918f-4ccb6b5bb2eb\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.960029 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6352d300-c932-4544-918f-4ccb6b5bb2eb-secret-volume\") pod \"6352d300-c932-4544-918f-4ccb6b5bb2eb\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.960072 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjvmm\" (UniqueName: \"kubernetes.io/projected/6352d300-c932-4544-918f-4ccb6b5bb2eb-kube-api-access-sjvmm\") pod \"6352d300-c932-4544-918f-4ccb6b5bb2eb\" (UID: \"6352d300-c932-4544-918f-4ccb6b5bb2eb\") " Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.961031 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6352d300-c932-4544-918f-4ccb6b5bb2eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "6352d300-c932-4544-918f-4ccb6b5bb2eb" (UID: "6352d300-c932-4544-918f-4ccb6b5bb2eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.966662 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6352d300-c932-4544-918f-4ccb6b5bb2eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6352d300-c932-4544-918f-4ccb6b5bb2eb" (UID: "6352d300-c932-4544-918f-4ccb6b5bb2eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:45:02 crc kubenswrapper[4796]: I0930 16:45:02.968447 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6352d300-c932-4544-918f-4ccb6b5bb2eb-kube-api-access-sjvmm" (OuterVolumeSpecName: "kube-api-access-sjvmm") pod "6352d300-c932-4544-918f-4ccb6b5bb2eb" (UID: "6352d300-c932-4544-918f-4ccb6b5bb2eb"). InnerVolumeSpecName "kube-api-access-sjvmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.062077 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6352d300-c932-4544-918f-4ccb6b5bb2eb-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.062132 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjvmm\" (UniqueName: \"kubernetes.io/projected/6352d300-c932-4544-918f-4ccb6b5bb2eb-kube-api-access-sjvmm\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.062144 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6352d300-c932-4544-918f-4ccb6b5bb2eb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.474315 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" event={"ID":"6352d300-c932-4544-918f-4ccb6b5bb2eb","Type":"ContainerDied","Data":"e17a51e768534c0cd1d8301bc9ba35b55da587e6be3b9500490bc4682f37c687"} Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.474829 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e17a51e768534c0cd1d8301bc9ba35b55da587e6be3b9500490bc4682f37c687" Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.474383 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320845-qnpf6" Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.885465 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z"] Sep 30 16:45:03 crc kubenswrapper[4796]: I0930 16:45:03.892800 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320800-pdt4z"] Sep 30 16:45:04 crc kubenswrapper[4796]: I0930 16:45:04.749212 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2528392f-8a0a-4ec7-8360-5c4f99dc14fa" path="/var/lib/kubelet/pods/2528392f-8a0a-4ec7-8360-5c4f99dc14fa/volumes" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.235879 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fxgjm"] Sep 30 16:45:10 crc kubenswrapper[4796]: E0930 16:45:10.237719 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6352d300-c932-4544-918f-4ccb6b5bb2eb" containerName="collect-profiles" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.237797 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6352d300-c932-4544-918f-4ccb6b5bb2eb" containerName="collect-profiles" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.238680 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6352d300-c932-4544-918f-4ccb6b5bb2eb" containerName="collect-profiles" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.240779 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.248684 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxgjm"] Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.411343 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-utilities\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.411771 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-catalog-content\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.411830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhtv8\" (UniqueName: \"kubernetes.io/projected/f58b0c2a-926a-4fef-93d4-6e8edd428603-kube-api-access-lhtv8\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.513822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-utilities\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.514044 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-catalog-content\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.514120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhtv8\" (UniqueName: \"kubernetes.io/projected/f58b0c2a-926a-4fef-93d4-6e8edd428603-kube-api-access-lhtv8\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.514654 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-catalog-content\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.514754 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-utilities\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.565756 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhtv8\" (UniqueName: \"kubernetes.io/projected/f58b0c2a-926a-4fef-93d4-6e8edd428603-kube-api-access-lhtv8\") pod \"community-operators-fxgjm\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:10 crc kubenswrapper[4796]: I0930 16:45:10.579799 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:11 crc kubenswrapper[4796]: I0930 16:45:11.104122 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxgjm"] Sep 30 16:45:11 crc kubenswrapper[4796]: E0930 16:45:11.526822 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf58b0c2a_926a_4fef_93d4_6e8edd428603.slice/crio-fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf58b0c2a_926a_4fef_93d4_6e8edd428603.slice/crio-conmon-fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:45:11 crc kubenswrapper[4796]: I0930 16:45:11.578534 4796 generic.go:334] "Generic (PLEG): container finished" podID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerID="fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae" exitCode=0 Sep 30 16:45:11 crc kubenswrapper[4796]: I0930 16:45:11.578592 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerDied","Data":"fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae"} Sep 30 16:45:11 crc kubenswrapper[4796]: I0930 16:45:11.578622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerStarted","Data":"47969352865a9467346da860112996b36beaaaaa43f01b4e0366caa1268bde7e"} Sep 30 16:45:12 crc kubenswrapper[4796]: I0930 16:45:12.588236 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerStarted","Data":"7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9"} Sep 30 16:45:13 crc kubenswrapper[4796]: I0930 16:45:13.603177 4796 generic.go:334] "Generic (PLEG): container finished" podID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerID="7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9" exitCode=0 Sep 30 16:45:13 crc kubenswrapper[4796]: I0930 16:45:13.603256 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerDied","Data":"7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9"} Sep 30 16:45:15 crc kubenswrapper[4796]: I0930 16:45:15.627696 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerStarted","Data":"0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433"} Sep 30 16:45:15 crc kubenswrapper[4796]: I0930 16:45:15.671411 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fxgjm" podStartSLOduration=2.7441780529999997 podStartE2EDuration="5.671387716s" podCreationTimestamp="2025-09-30 16:45:10 +0000 UTC" firstStartedPulling="2025-09-30 16:45:11.581456677 +0000 UTC m=+2003.594735214" lastFinishedPulling="2025-09-30 16:45:14.50866631 +0000 UTC m=+2006.521944877" observedRunningTime="2025-09-30 16:45:15.651128551 +0000 UTC m=+2007.664407128" watchObservedRunningTime="2025-09-30 16:45:15.671387716 +0000 UTC m=+2007.684666253" Sep 30 16:45:20 crc kubenswrapper[4796]: I0930 16:45:20.580381 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:20 crc kubenswrapper[4796]: I0930 16:45:20.580959 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:20 crc kubenswrapper[4796]: I0930 16:45:20.633577 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:20 crc kubenswrapper[4796]: I0930 16:45:20.722218 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:20 crc kubenswrapper[4796]: I0930 16:45:20.874101 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxgjm"] Sep 30 16:45:22 crc kubenswrapper[4796]: I0930 16:45:22.690574 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fxgjm" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="registry-server" containerID="cri-o://0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433" gracePeriod=2 Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.120863 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.274242 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhtv8\" (UniqueName: \"kubernetes.io/projected/f58b0c2a-926a-4fef-93d4-6e8edd428603-kube-api-access-lhtv8\") pod \"f58b0c2a-926a-4fef-93d4-6e8edd428603\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.274319 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-catalog-content\") pod \"f58b0c2a-926a-4fef-93d4-6e8edd428603\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.274585 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-utilities\") pod \"f58b0c2a-926a-4fef-93d4-6e8edd428603\" (UID: \"f58b0c2a-926a-4fef-93d4-6e8edd428603\") " Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.275413 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-utilities" (OuterVolumeSpecName: "utilities") pod "f58b0c2a-926a-4fef-93d4-6e8edd428603" (UID: "f58b0c2a-926a-4fef-93d4-6e8edd428603"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.281604 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f58b0c2a-926a-4fef-93d4-6e8edd428603-kube-api-access-lhtv8" (OuterVolumeSpecName: "kube-api-access-lhtv8") pod "f58b0c2a-926a-4fef-93d4-6e8edd428603" (UID: "f58b0c2a-926a-4fef-93d4-6e8edd428603"). InnerVolumeSpecName "kube-api-access-lhtv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.354672 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f58b0c2a-926a-4fef-93d4-6e8edd428603" (UID: "f58b0c2a-926a-4fef-93d4-6e8edd428603"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.376786 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.376835 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhtv8\" (UniqueName: \"kubernetes.io/projected/f58b0c2a-926a-4fef-93d4-6e8edd428603-kube-api-access-lhtv8\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.376850 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f58b0c2a-926a-4fef-93d4-6e8edd428603-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.707262 4796 generic.go:334] "Generic (PLEG): container finished" podID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerID="0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433" exitCode=0 Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.707326 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerDied","Data":"0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433"} Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.707354 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxgjm" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.707381 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxgjm" event={"ID":"f58b0c2a-926a-4fef-93d4-6e8edd428603","Type":"ContainerDied","Data":"47969352865a9467346da860112996b36beaaaaa43f01b4e0366caa1268bde7e"} Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.707412 4796 scope.go:117] "RemoveContainer" containerID="0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.740795 4796 scope.go:117] "RemoveContainer" containerID="7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.780090 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxgjm"] Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.788507 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fxgjm"] Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.813282 4796 scope.go:117] "RemoveContainer" containerID="fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.843746 4796 scope.go:117] "RemoveContainer" containerID="0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433" Sep 30 16:45:23 crc kubenswrapper[4796]: E0930 16:45:23.844368 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433\": container with ID starting with 0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433 not found: ID does not exist" containerID="0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.844443 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433"} err="failed to get container status \"0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433\": rpc error: code = NotFound desc = could not find container \"0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433\": container with ID starting with 0dbbfa69c85a922650b209d8bd57d44799e6274f17483354d27bdee8821fd433 not found: ID does not exist" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.844507 4796 scope.go:117] "RemoveContainer" containerID="7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9" Sep 30 16:45:23 crc kubenswrapper[4796]: E0930 16:45:23.844947 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9\": container with ID starting with 7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9 not found: ID does not exist" containerID="7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.845019 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9"} err="failed to get container status \"7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9\": rpc error: code = NotFound desc = could not find container \"7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9\": container with ID starting with 7e8048038b3cb59bdf6f1b789500515b901f33fcddd0c9dca7c6b63a3df4fbf9 not found: ID does not exist" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.845045 4796 scope.go:117] "RemoveContainer" containerID="fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae" Sep 30 16:45:23 crc kubenswrapper[4796]: E0930 16:45:23.845334 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae\": container with ID starting with fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae not found: ID does not exist" containerID="fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae" Sep 30 16:45:23 crc kubenswrapper[4796]: I0930 16:45:23.845371 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae"} err="failed to get container status \"fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae\": rpc error: code = NotFound desc = could not find container \"fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae\": container with ID starting with fd450f3785021409664a02bf12c9a625c3d74fc76d7a41eb3bdc5245cb3a98ae not found: ID does not exist" Sep 30 16:45:24 crc kubenswrapper[4796]: I0930 16:45:24.747656 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" path="/var/lib/kubelet/pods/f58b0c2a-926a-4fef-93d4-6e8edd428603/volumes" Sep 30 16:45:40 crc kubenswrapper[4796]: I0930 16:45:40.888766 4796 generic.go:334] "Generic (PLEG): container finished" podID="abb53121-c532-478a-9670-8350dd3e477b" containerID="6403599bedb94f41bf83ded6db807312a66aa633747940ba3173fe5f69db1673" exitCode=0 Sep 30 16:45:40 crc kubenswrapper[4796]: I0930 16:45:40.888839 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" event={"ID":"abb53121-c532-478a-9670-8350dd3e477b","Type":"ContainerDied","Data":"6403599bedb94f41bf83ded6db807312a66aa633747940ba3173fe5f69db1673"} Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.352669 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.492014 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg8xs\" (UniqueName: \"kubernetes.io/projected/abb53121-c532-478a-9670-8350dd3e477b-kube-api-access-bg8xs\") pod \"abb53121-c532-478a-9670-8350dd3e477b\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.492101 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/abb53121-c532-478a-9670-8350dd3e477b-ovncontroller-config-0\") pod \"abb53121-c532-478a-9670-8350dd3e477b\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.492144 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ssh-key\") pod \"abb53121-c532-478a-9670-8350dd3e477b\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.492268 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-inventory\") pod \"abb53121-c532-478a-9670-8350dd3e477b\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.492335 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ovn-combined-ca-bundle\") pod \"abb53121-c532-478a-9670-8350dd3e477b\" (UID: \"abb53121-c532-478a-9670-8350dd3e477b\") " Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.523605 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "abb53121-c532-478a-9670-8350dd3e477b" (UID: "abb53121-c532-478a-9670-8350dd3e477b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.523669 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb53121-c532-478a-9670-8350dd3e477b-kube-api-access-bg8xs" (OuterVolumeSpecName: "kube-api-access-bg8xs") pod "abb53121-c532-478a-9670-8350dd3e477b" (UID: "abb53121-c532-478a-9670-8350dd3e477b"). InnerVolumeSpecName "kube-api-access-bg8xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.525529 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abb53121-c532-478a-9670-8350dd3e477b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "abb53121-c532-478a-9670-8350dd3e477b" (UID: "abb53121-c532-478a-9670-8350dd3e477b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.527735 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-inventory" (OuterVolumeSpecName: "inventory") pod "abb53121-c532-478a-9670-8350dd3e477b" (UID: "abb53121-c532-478a-9670-8350dd3e477b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.529587 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abb53121-c532-478a-9670-8350dd3e477b" (UID: "abb53121-c532-478a-9670-8350dd3e477b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.594971 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.595013 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.595024 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg8xs\" (UniqueName: \"kubernetes.io/projected/abb53121-c532-478a-9670-8350dd3e477b-kube-api-access-bg8xs\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.595035 4796 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/abb53121-c532-478a-9670-8350dd3e477b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.595045 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abb53121-c532-478a-9670-8350dd3e477b-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.912950 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" event={"ID":"abb53121-c532-478a-9670-8350dd3e477b","Type":"ContainerDied","Data":"a210593a4a3c83f342c2bb7a928dd68e60e92c1d5b38f01a8dcfd1b92baced06"} Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.913284 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a210593a4a3c83f342c2bb7a928dd68e60e92c1d5b38f01a8dcfd1b92baced06" Sep 30 16:45:42 crc kubenswrapper[4796]: I0930 16:45:42.913044 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vtt88" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.023704 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds"] Sep 30 16:45:43 crc kubenswrapper[4796]: E0930 16:45:43.024303 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb53121-c532-478a-9670-8350dd3e477b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.024349 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb53121-c532-478a-9670-8350dd3e477b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 16:45:43 crc kubenswrapper[4796]: E0930 16:45:43.024392 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="registry-server" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.024404 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="registry-server" Sep 30 16:45:43 crc kubenswrapper[4796]: E0930 16:45:43.024446 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="extract-utilities" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.024458 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="extract-utilities" Sep 30 16:45:43 crc kubenswrapper[4796]: E0930 16:45:43.024477 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="extract-content" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.024486 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="extract-content" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.024820 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb53121-c532-478a-9670-8350dd3e477b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.024862 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f58b0c2a-926a-4fef-93d4-6e8edd428603" containerName="registry-server" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.025884 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.027954 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.028296 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.029844 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.030155 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.030293 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.035018 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.037418 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds"] Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.107340 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.107405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkhrh\" (UniqueName: \"kubernetes.io/projected/f90058d7-bebf-4a04-8fb5-045010f7a6b3-kube-api-access-kkhrh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.107562 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.107670 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.107717 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.107753 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.209502 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.209554 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.209581 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.209611 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.209634 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkhrh\" (UniqueName: \"kubernetes.io/projected/f90058d7-bebf-4a04-8fb5-045010f7a6b3-kube-api-access-kkhrh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.209701 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.214309 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.215475 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.215588 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.215609 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.216464 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.231540 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkhrh\" (UniqueName: \"kubernetes.io/projected/f90058d7-bebf-4a04-8fb5-045010f7a6b3-kube-api-access-kkhrh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.351769 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.920561 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds"] Sep 30 16:45:43 crc kubenswrapper[4796]: I0930 16:45:43.930175 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" event={"ID":"f90058d7-bebf-4a04-8fb5-045010f7a6b3","Type":"ContainerStarted","Data":"d239b6cfcb9e545baf0ad406f3d93a2d5e6ac61d29ccf036722247b8a20d44ea"} Sep 30 16:45:44 crc kubenswrapper[4796]: I0930 16:45:44.943105 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" event={"ID":"f90058d7-bebf-4a04-8fb5-045010f7a6b3","Type":"ContainerStarted","Data":"302590a017ca0448c7f79c9fca353d32fc387b1608cf9ac1f5fcd9b36cbf97e5"} Sep 30 16:45:44 crc kubenswrapper[4796]: I0930 16:45:44.977879 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" podStartSLOduration=2.435212379 podStartE2EDuration="2.977856895s" podCreationTimestamp="2025-09-30 16:45:42 +0000 UTC" firstStartedPulling="2025-09-30 16:45:43.922395883 +0000 UTC m=+2035.935674420" lastFinishedPulling="2025-09-30 16:45:44.465040399 +0000 UTC m=+2036.478318936" observedRunningTime="2025-09-30 16:45:44.961409401 +0000 UTC m=+2036.974687948" watchObservedRunningTime="2025-09-30 16:45:44.977856895 +0000 UTC m=+2036.991135422" Sep 30 16:45:49 crc kubenswrapper[4796]: I0930 16:45:49.875192 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-trwmj"] Sep 30 16:45:49 crc kubenswrapper[4796]: I0930 16:45:49.891829 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-trwmj"] Sep 30 16:45:49 crc kubenswrapper[4796]: I0930 16:45:49.891968 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:49 crc kubenswrapper[4796]: I0930 16:45:49.963290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kfvw\" (UniqueName: \"kubernetes.io/projected/a9aacf33-35ba-45dc-a451-d27f953760fd-kube-api-access-2kfvw\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:49 crc kubenswrapper[4796]: I0930 16:45:49.963335 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-utilities\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:49 crc kubenswrapper[4796]: I0930 16:45:49.963488 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-catalog-content\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.065016 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-catalog-content\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.065085 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kfvw\" (UniqueName: \"kubernetes.io/projected/a9aacf33-35ba-45dc-a451-d27f953760fd-kube-api-access-2kfvw\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.065124 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-utilities\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.065695 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-catalog-content\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.066020 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-utilities\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.084579 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kfvw\" (UniqueName: \"kubernetes.io/projected/a9aacf33-35ba-45dc-a451-d27f953760fd-kube-api-access-2kfvw\") pod \"redhat-operators-trwmj\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.217666 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:45:50 crc kubenswrapper[4796]: I0930 16:45:50.683041 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-trwmj"] Sep 30 16:45:51 crc kubenswrapper[4796]: I0930 16:45:51.004777 4796 generic.go:334] "Generic (PLEG): container finished" podID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerID="7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827" exitCode=0 Sep 30 16:45:51 crc kubenswrapper[4796]: I0930 16:45:51.004883 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trwmj" event={"ID":"a9aacf33-35ba-45dc-a451-d27f953760fd","Type":"ContainerDied","Data":"7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827"} Sep 30 16:45:51 crc kubenswrapper[4796]: I0930 16:45:51.005126 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trwmj" event={"ID":"a9aacf33-35ba-45dc-a451-d27f953760fd","Type":"ContainerStarted","Data":"9bd3e132ff50520cc70962a1b2604a93a64cf94a912adabb45831eda1c69ce42"} Sep 30 16:45:53 crc kubenswrapper[4796]: I0930 16:45:53.027518 4796 generic.go:334] "Generic (PLEG): container finished" podID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerID="4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890" exitCode=0 Sep 30 16:45:53 crc kubenswrapper[4796]: I0930 16:45:53.027716 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trwmj" event={"ID":"a9aacf33-35ba-45dc-a451-d27f953760fd","Type":"ContainerDied","Data":"4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890"} Sep 30 16:45:54 crc kubenswrapper[4796]: I0930 16:45:54.048318 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trwmj" event={"ID":"a9aacf33-35ba-45dc-a451-d27f953760fd","Type":"ContainerStarted","Data":"f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e"} Sep 30 16:45:54 crc kubenswrapper[4796]: I0930 16:45:54.073849 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-trwmj" podStartSLOduration=2.628882261 podStartE2EDuration="5.073823008s" podCreationTimestamp="2025-09-30 16:45:49 +0000 UTC" firstStartedPulling="2025-09-30 16:45:51.006102724 +0000 UTC m=+2043.019381251" lastFinishedPulling="2025-09-30 16:45:53.451043471 +0000 UTC m=+2045.464321998" observedRunningTime="2025-09-30 16:45:54.069576206 +0000 UTC m=+2046.082854743" watchObservedRunningTime="2025-09-30 16:45:54.073823008 +0000 UTC m=+2046.087101565" Sep 30 16:45:55 crc kubenswrapper[4796]: I0930 16:45:55.476888 4796 scope.go:117] "RemoveContainer" containerID="4466cdef91302f92d7fc3272496856720d0368f03a17525d1e0decebe57ebdd8" Sep 30 16:46:00 crc kubenswrapper[4796]: I0930 16:46:00.218723 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:46:00 crc kubenswrapper[4796]: I0930 16:46:00.219318 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:46:00 crc kubenswrapper[4796]: I0930 16:46:00.268369 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:46:01 crc kubenswrapper[4796]: I0930 16:46:01.180881 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:46:01 crc kubenswrapper[4796]: I0930 16:46:01.227459 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-trwmj"] Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.143490 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-trwmj" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="registry-server" containerID="cri-o://f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e" gracePeriod=2 Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.582570 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.646156 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kfvw\" (UniqueName: \"kubernetes.io/projected/a9aacf33-35ba-45dc-a451-d27f953760fd-kube-api-access-2kfvw\") pod \"a9aacf33-35ba-45dc-a451-d27f953760fd\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.646228 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-catalog-content\") pod \"a9aacf33-35ba-45dc-a451-d27f953760fd\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.646345 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-utilities\") pod \"a9aacf33-35ba-45dc-a451-d27f953760fd\" (UID: \"a9aacf33-35ba-45dc-a451-d27f953760fd\") " Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.648079 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-utilities" (OuterVolumeSpecName: "utilities") pod "a9aacf33-35ba-45dc-a451-d27f953760fd" (UID: "a9aacf33-35ba-45dc-a451-d27f953760fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.653302 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9aacf33-35ba-45dc-a451-d27f953760fd-kube-api-access-2kfvw" (OuterVolumeSpecName: "kube-api-access-2kfvw") pod "a9aacf33-35ba-45dc-a451-d27f953760fd" (UID: "a9aacf33-35ba-45dc-a451-d27f953760fd"). InnerVolumeSpecName "kube-api-access-2kfvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.735339 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9aacf33-35ba-45dc-a451-d27f953760fd" (UID: "a9aacf33-35ba-45dc-a451-d27f953760fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.750797 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.750855 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kfvw\" (UniqueName: \"kubernetes.io/projected/a9aacf33-35ba-45dc-a451-d27f953760fd-kube-api-access-2kfvw\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:03 crc kubenswrapper[4796]: I0930 16:46:03.750872 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9aacf33-35ba-45dc-a451-d27f953760fd-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.157309 4796 generic.go:334] "Generic (PLEG): container finished" podID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerID="f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e" exitCode=0 Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.157402 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trwmj" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.157364 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trwmj" event={"ID":"a9aacf33-35ba-45dc-a451-d27f953760fd","Type":"ContainerDied","Data":"f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e"} Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.157636 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trwmj" event={"ID":"a9aacf33-35ba-45dc-a451-d27f953760fd","Type":"ContainerDied","Data":"9bd3e132ff50520cc70962a1b2604a93a64cf94a912adabb45831eda1c69ce42"} Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.157773 4796 scope.go:117] "RemoveContainer" containerID="f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.193132 4796 scope.go:117] "RemoveContainer" containerID="4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.200164 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-trwmj"] Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.204957 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-trwmj"] Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.252677 4796 scope.go:117] "RemoveContainer" containerID="7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.304347 4796 scope.go:117] "RemoveContainer" containerID="f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e" Sep 30 16:46:04 crc kubenswrapper[4796]: E0930 16:46:04.305014 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e\": container with ID starting with f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e not found: ID does not exist" containerID="f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.305058 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e"} err="failed to get container status \"f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e\": rpc error: code = NotFound desc = could not find container \"f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e\": container with ID starting with f935bf6acf95523a540d1f7514c081a85e1747fbd9f8cd473e1044fbc0706e4e not found: ID does not exist" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.305087 4796 scope.go:117] "RemoveContainer" containerID="4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890" Sep 30 16:46:04 crc kubenswrapper[4796]: E0930 16:46:04.305407 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890\": container with ID starting with 4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890 not found: ID does not exist" containerID="4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.305461 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890"} err="failed to get container status \"4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890\": rpc error: code = NotFound desc = could not find container \"4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890\": container with ID starting with 4cf7fbac7150be1e54d590407f64d518bd9a160c4e26263b379246ad437b2890 not found: ID does not exist" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.305497 4796 scope.go:117] "RemoveContainer" containerID="7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827" Sep 30 16:46:04 crc kubenswrapper[4796]: E0930 16:46:04.305782 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827\": container with ID starting with 7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827 not found: ID does not exist" containerID="7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.305926 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827"} err="failed to get container status \"7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827\": rpc error: code = NotFound desc = could not find container \"7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827\": container with ID starting with 7b73a5d09d100a3827731240fc503ae06a73d74a17b52a4f92cd177a30d49827 not found: ID does not exist" Sep 30 16:46:04 crc kubenswrapper[4796]: I0930 16:46:04.747442 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" path="/var/lib/kubelet/pods/a9aacf33-35ba-45dc-a451-d27f953760fd/volumes" Sep 30 16:46:39 crc kubenswrapper[4796]: I0930 16:46:39.549831 4796 generic.go:334] "Generic (PLEG): container finished" podID="f90058d7-bebf-4a04-8fb5-045010f7a6b3" containerID="302590a017ca0448c7f79c9fca353d32fc387b1608cf9ac1f5fcd9b36cbf97e5" exitCode=0 Sep 30 16:46:39 crc kubenswrapper[4796]: I0930 16:46:39.549953 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" event={"ID":"f90058d7-bebf-4a04-8fb5-045010f7a6b3","Type":"ContainerDied","Data":"302590a017ca0448c7f79c9fca353d32fc387b1608cf9ac1f5fcd9b36cbf97e5"} Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.058250 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.102785 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-ssh-key\") pod \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.102891 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.103009 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-metadata-combined-ca-bundle\") pod \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.103126 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-inventory\") pod \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.103874 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-nova-metadata-neutron-config-0\") pod \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.103961 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkhrh\" (UniqueName: \"kubernetes.io/projected/f90058d7-bebf-4a04-8fb5-045010f7a6b3-kube-api-access-kkhrh\") pod \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\" (UID: \"f90058d7-bebf-4a04-8fb5-045010f7a6b3\") " Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.109733 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f90058d7-bebf-4a04-8fb5-045010f7a6b3" (UID: "f90058d7-bebf-4a04-8fb5-045010f7a6b3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.111395 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90058d7-bebf-4a04-8fb5-045010f7a6b3-kube-api-access-kkhrh" (OuterVolumeSpecName: "kube-api-access-kkhrh") pod "f90058d7-bebf-4a04-8fb5-045010f7a6b3" (UID: "f90058d7-bebf-4a04-8fb5-045010f7a6b3"). InnerVolumeSpecName "kube-api-access-kkhrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.132143 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "f90058d7-bebf-4a04-8fb5-045010f7a6b3" (UID: "f90058d7-bebf-4a04-8fb5-045010f7a6b3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.136539 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f90058d7-bebf-4a04-8fb5-045010f7a6b3" (UID: "f90058d7-bebf-4a04-8fb5-045010f7a6b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.151379 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-inventory" (OuterVolumeSpecName: "inventory") pod "f90058d7-bebf-4a04-8fb5-045010f7a6b3" (UID: "f90058d7-bebf-4a04-8fb5-045010f7a6b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.157745 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "f90058d7-bebf-4a04-8fb5-045010f7a6b3" (UID: "f90058d7-bebf-4a04-8fb5-045010f7a6b3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.210534 4796 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.210868 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.210959 4796 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.211112 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkhrh\" (UniqueName: \"kubernetes.io/projected/f90058d7-bebf-4a04-8fb5-045010f7a6b3-kube-api-access-kkhrh\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.211207 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.211280 4796 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f90058d7-bebf-4a04-8fb5-045010f7a6b3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.580793 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" event={"ID":"f90058d7-bebf-4a04-8fb5-045010f7a6b3","Type":"ContainerDied","Data":"d239b6cfcb9e545baf0ad406f3d93a2d5e6ac61d29ccf036722247b8a20d44ea"} Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.580861 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d239b6cfcb9e545baf0ad406f3d93a2d5e6ac61d29ccf036722247b8a20d44ea" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.580907 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.777521 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx"] Sep 30 16:46:41 crc kubenswrapper[4796]: E0930 16:46:41.778670 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90058d7-bebf-4a04-8fb5-045010f7a6b3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.778806 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90058d7-bebf-4a04-8fb5-045010f7a6b3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 16:46:41 crc kubenswrapper[4796]: E0930 16:46:41.778951 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="extract-content" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.779094 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="extract-content" Sep 30 16:46:41 crc kubenswrapper[4796]: E0930 16:46:41.779321 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="extract-utilities" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.779460 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="extract-utilities" Sep 30 16:46:41 crc kubenswrapper[4796]: E0930 16:46:41.779596 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="registry-server" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.779684 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="registry-server" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.780116 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90058d7-bebf-4a04-8fb5-045010f7a6b3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.780228 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9aacf33-35ba-45dc-a451-d27f953760fd" containerName="registry-server" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.781116 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.784105 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.784357 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.784565 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.784623 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.784581 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.791067 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx"] Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.827430 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.828132 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4lvw\" (UniqueName: \"kubernetes.io/projected/0f40019c-f145-46f4-89f3-cc7244d69faf-kube-api-access-b4lvw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.828243 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.828373 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.829195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.930851 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.931008 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.931169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.931303 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.931470 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4lvw\" (UniqueName: \"kubernetes.io/projected/0f40019c-f145-46f4-89f3-cc7244d69faf-kube-api-access-b4lvw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.938458 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.939301 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.940069 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.943384 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:41 crc kubenswrapper[4796]: I0930 16:46:41.952933 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4lvw\" (UniqueName: \"kubernetes.io/projected/0f40019c-f145-46f4-89f3-cc7244d69faf-kube-api-access-b4lvw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-89pdx\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:42 crc kubenswrapper[4796]: I0930 16:46:42.111557 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:46:42 crc kubenswrapper[4796]: I0930 16:46:42.694826 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx"] Sep 30 16:46:43 crc kubenswrapper[4796]: I0930 16:46:43.607763 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" event={"ID":"0f40019c-f145-46f4-89f3-cc7244d69faf","Type":"ContainerStarted","Data":"2f04d29b4e5e606156ec0d6e0b2a07e3461841e1c789ff6d27924a46815f5973"} Sep 30 16:46:44 crc kubenswrapper[4796]: I0930 16:46:44.640587 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" event={"ID":"0f40019c-f145-46f4-89f3-cc7244d69faf","Type":"ContainerStarted","Data":"b096567650da35b005e4f38d720af409f20acf1c5c5b344ba4f26a3c1b4ea5e9"} Sep 30 16:46:44 crc kubenswrapper[4796]: I0930 16:46:44.664052 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" podStartSLOduration=2.715416207 podStartE2EDuration="3.664034289s" podCreationTimestamp="2025-09-30 16:46:41 +0000 UTC" firstStartedPulling="2025-09-30 16:46:42.701289915 +0000 UTC m=+2094.714568482" lastFinishedPulling="2025-09-30 16:46:43.649907997 +0000 UTC m=+2095.663186564" observedRunningTime="2025-09-30 16:46:44.661068463 +0000 UTC m=+2096.674346990" watchObservedRunningTime="2025-09-30 16:46:44.664034289 +0000 UTC m=+2096.677312806" Sep 30 16:46:52 crc kubenswrapper[4796]: I0930 16:46:52.869405 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d67r2"] Sep 30 16:46:52 crc kubenswrapper[4796]: I0930 16:46:52.871963 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:52 crc kubenswrapper[4796]: I0930 16:46:52.905657 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d67r2"] Sep 30 16:46:52 crc kubenswrapper[4796]: I0930 16:46:52.965038 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnrzj\" (UniqueName: \"kubernetes.io/projected/96334f92-e1e6-4261-bc5d-b2310640a925-kube-api-access-tnrzj\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:52 crc kubenswrapper[4796]: I0930 16:46:52.965079 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-utilities\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:52 crc kubenswrapper[4796]: I0930 16:46:52.965346 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-catalog-content\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.067345 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-utilities\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.067611 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnrzj\" (UniqueName: \"kubernetes.io/projected/96334f92-e1e6-4261-bc5d-b2310640a925-kube-api-access-tnrzj\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.067734 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-catalog-content\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.068384 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-catalog-content\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.068381 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-utilities\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.089959 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnrzj\" (UniqueName: \"kubernetes.io/projected/96334f92-e1e6-4261-bc5d-b2310640a925-kube-api-access-tnrzj\") pod \"redhat-marketplace-d67r2\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.203530 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.698484 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d67r2"] Sep 30 16:46:53 crc kubenswrapper[4796]: I0930 16:46:53.749240 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d67r2" event={"ID":"96334f92-e1e6-4261-bc5d-b2310640a925","Type":"ContainerStarted","Data":"48d5f3ff9f4b3b945603b86e80fae6e02f3eef60cc460aabeb0710e2915576db"} Sep 30 16:46:54 crc kubenswrapper[4796]: E0930 16:46:54.186073 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96334f92_e1e6_4261_bc5d_b2310640a925.slice/crio-b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d.scope\": RecentStats: unable to find data in memory cache]" Sep 30 16:46:54 crc kubenswrapper[4796]: I0930 16:46:54.763373 4796 generic.go:334] "Generic (PLEG): container finished" podID="96334f92-e1e6-4261-bc5d-b2310640a925" containerID="b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d" exitCode=0 Sep 30 16:46:54 crc kubenswrapper[4796]: I0930 16:46:54.763517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d67r2" event={"ID":"96334f92-e1e6-4261-bc5d-b2310640a925","Type":"ContainerDied","Data":"b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d"} Sep 30 16:46:56 crc kubenswrapper[4796]: I0930 16:46:56.805170 4796 generic.go:334] "Generic (PLEG): container finished" podID="96334f92-e1e6-4261-bc5d-b2310640a925" containerID="715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d" exitCode=0 Sep 30 16:46:56 crc kubenswrapper[4796]: I0930 16:46:56.805292 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d67r2" event={"ID":"96334f92-e1e6-4261-bc5d-b2310640a925","Type":"ContainerDied","Data":"715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d"} Sep 30 16:46:57 crc kubenswrapper[4796]: I0930 16:46:57.818883 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d67r2" event={"ID":"96334f92-e1e6-4261-bc5d-b2310640a925","Type":"ContainerStarted","Data":"93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f"} Sep 30 16:46:57 crc kubenswrapper[4796]: I0930 16:46:57.853285 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d67r2" podStartSLOduration=3.389419493 podStartE2EDuration="5.853255515s" podCreationTimestamp="2025-09-30 16:46:52 +0000 UTC" firstStartedPulling="2025-09-30 16:46:54.76678604 +0000 UTC m=+2106.780064597" lastFinishedPulling="2025-09-30 16:46:57.230622052 +0000 UTC m=+2109.243900619" observedRunningTime="2025-09-30 16:46:57.844640957 +0000 UTC m=+2109.857919494" watchObservedRunningTime="2025-09-30 16:46:57.853255515 +0000 UTC m=+2109.866534052" Sep 30 16:47:01 crc kubenswrapper[4796]: I0930 16:47:01.097789 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:47:01 crc kubenswrapper[4796]: I0930 16:47:01.098366 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:47:03 crc kubenswrapper[4796]: I0930 16:47:03.204329 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:47:03 crc kubenswrapper[4796]: I0930 16:47:03.204691 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:47:03 crc kubenswrapper[4796]: I0930 16:47:03.277909 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:47:03 crc kubenswrapper[4796]: I0930 16:47:03.948017 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:47:04 crc kubenswrapper[4796]: I0930 16:47:04.008708 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d67r2"] Sep 30 16:47:05 crc kubenswrapper[4796]: I0930 16:47:05.940135 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d67r2" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="registry-server" containerID="cri-o://93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f" gracePeriod=2 Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.398129 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.455116 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnrzj\" (UniqueName: \"kubernetes.io/projected/96334f92-e1e6-4261-bc5d-b2310640a925-kube-api-access-tnrzj\") pod \"96334f92-e1e6-4261-bc5d-b2310640a925\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.455214 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-catalog-content\") pod \"96334f92-e1e6-4261-bc5d-b2310640a925\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.455354 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-utilities\") pod \"96334f92-e1e6-4261-bc5d-b2310640a925\" (UID: \"96334f92-e1e6-4261-bc5d-b2310640a925\") " Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.456620 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-utilities" (OuterVolumeSpecName: "utilities") pod "96334f92-e1e6-4261-bc5d-b2310640a925" (UID: "96334f92-e1e6-4261-bc5d-b2310640a925"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.461589 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96334f92-e1e6-4261-bc5d-b2310640a925-kube-api-access-tnrzj" (OuterVolumeSpecName: "kube-api-access-tnrzj") pod "96334f92-e1e6-4261-bc5d-b2310640a925" (UID: "96334f92-e1e6-4261-bc5d-b2310640a925"). InnerVolumeSpecName "kube-api-access-tnrzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.473959 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96334f92-e1e6-4261-bc5d-b2310640a925" (UID: "96334f92-e1e6-4261-bc5d-b2310640a925"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.557559 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.557795 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnrzj\" (UniqueName: \"kubernetes.io/projected/96334f92-e1e6-4261-bc5d-b2310640a925-kube-api-access-tnrzj\") on node \"crc\" DevicePath \"\"" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.557807 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96334f92-e1e6-4261-bc5d-b2310640a925-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.953194 4796 generic.go:334] "Generic (PLEG): container finished" podID="96334f92-e1e6-4261-bc5d-b2310640a925" containerID="93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f" exitCode=0 Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.953257 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d67r2" event={"ID":"96334f92-e1e6-4261-bc5d-b2310640a925","Type":"ContainerDied","Data":"93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f"} Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.953296 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d67r2" event={"ID":"96334f92-e1e6-4261-bc5d-b2310640a925","Type":"ContainerDied","Data":"48d5f3ff9f4b3b945603b86e80fae6e02f3eef60cc460aabeb0710e2915576db"} Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.953333 4796 scope.go:117] "RemoveContainer" containerID="93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.953517 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d67r2" Sep 30 16:47:06 crc kubenswrapper[4796]: I0930 16:47:06.994436 4796 scope.go:117] "RemoveContainer" containerID="715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.002728 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d67r2"] Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.015346 4796 scope.go:117] "RemoveContainer" containerID="b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.015971 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d67r2"] Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.060964 4796 scope.go:117] "RemoveContainer" containerID="93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f" Sep 30 16:47:07 crc kubenswrapper[4796]: E0930 16:47:07.061598 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f\": container with ID starting with 93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f not found: ID does not exist" containerID="93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.061655 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f"} err="failed to get container status \"93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f\": rpc error: code = NotFound desc = could not find container \"93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f\": container with ID starting with 93539a0f448bd955ce20b5ae8636bc6ca3e73a990c4f1959e7f24f5537cf413f not found: ID does not exist" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.061687 4796 scope.go:117] "RemoveContainer" containerID="715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d" Sep 30 16:47:07 crc kubenswrapper[4796]: E0930 16:47:07.062332 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d\": container with ID starting with 715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d not found: ID does not exist" containerID="715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.062362 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d"} err="failed to get container status \"715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d\": rpc error: code = NotFound desc = could not find container \"715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d\": container with ID starting with 715b78db4c7d0034d4ac17980b538a0729448ca80bb5d17a8d38d1548de7295d not found: ID does not exist" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.062380 4796 scope.go:117] "RemoveContainer" containerID="b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d" Sep 30 16:47:07 crc kubenswrapper[4796]: E0930 16:47:07.062601 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d\": container with ID starting with b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d not found: ID does not exist" containerID="b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d" Sep 30 16:47:07 crc kubenswrapper[4796]: I0930 16:47:07.062628 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d"} err="failed to get container status \"b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d\": rpc error: code = NotFound desc = could not find container \"b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d\": container with ID starting with b64a2e05030335996e55b06a77cbf691707b28453ee16f6a3f82e8b5e5d9b71d not found: ID does not exist" Sep 30 16:47:08 crc kubenswrapper[4796]: I0930 16:47:08.747936 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" path="/var/lib/kubelet/pods/96334f92-e1e6-4261-bc5d-b2310640a925/volumes" Sep 30 16:47:31 crc kubenswrapper[4796]: I0930 16:47:31.097747 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:47:31 crc kubenswrapper[4796]: I0930 16:47:31.098403 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.097160 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.097679 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.097727 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.098534 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.098619 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" gracePeriod=600 Sep 30 16:48:01 crc kubenswrapper[4796]: E0930 16:48:01.233745 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.563889 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" exitCode=0 Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.563957 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752"} Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.564040 4796 scope.go:117] "RemoveContainer" containerID="19814df5d3863819d92e9dc4e7cf4384be1a02052dcbbc67f85902a6c3e2db5c" Sep 30 16:48:01 crc kubenswrapper[4796]: I0930 16:48:01.565052 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:48:01 crc kubenswrapper[4796]: E0930 16:48:01.565591 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:48:14 crc kubenswrapper[4796]: I0930 16:48:14.734835 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:48:14 crc kubenswrapper[4796]: E0930 16:48:14.736211 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:48:27 crc kubenswrapper[4796]: I0930 16:48:27.733428 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:48:27 crc kubenswrapper[4796]: E0930 16:48:27.734395 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:48:41 crc kubenswrapper[4796]: I0930 16:48:41.733745 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:48:41 crc kubenswrapper[4796]: E0930 16:48:41.734556 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:48:54 crc kubenswrapper[4796]: I0930 16:48:54.733686 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:48:54 crc kubenswrapper[4796]: E0930 16:48:54.734739 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:49:05 crc kubenswrapper[4796]: I0930 16:49:05.733944 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:49:05 crc kubenswrapper[4796]: E0930 16:49:05.735210 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:49:20 crc kubenswrapper[4796]: I0930 16:49:20.733693 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:49:20 crc kubenswrapper[4796]: E0930 16:49:20.734413 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:49:35 crc kubenswrapper[4796]: I0930 16:49:35.733156 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:49:35 crc kubenswrapper[4796]: E0930 16:49:35.734177 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:49:46 crc kubenswrapper[4796]: I0930 16:49:46.734178 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:49:46 crc kubenswrapper[4796]: E0930 16:49:46.735330 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:49:58 crc kubenswrapper[4796]: I0930 16:49:58.752006 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:49:58 crc kubenswrapper[4796]: E0930 16:49:58.753053 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:50:12 crc kubenswrapper[4796]: I0930 16:50:12.735062 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:50:12 crc kubenswrapper[4796]: E0930 16:50:12.737300 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:50:24 crc kubenswrapper[4796]: I0930 16:50:24.734093 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:50:24 crc kubenswrapper[4796]: E0930 16:50:24.735449 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:50:37 crc kubenswrapper[4796]: I0930 16:50:37.733605 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:50:37 crc kubenswrapper[4796]: E0930 16:50:37.734164 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:50:52 crc kubenswrapper[4796]: I0930 16:50:52.734139 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:50:52 crc kubenswrapper[4796]: E0930 16:50:52.735260 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:51:06 crc kubenswrapper[4796]: I0930 16:51:06.733472 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:51:06 crc kubenswrapper[4796]: E0930 16:51:06.734390 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:51:20 crc kubenswrapper[4796]: I0930 16:51:20.733882 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:51:20 crc kubenswrapper[4796]: E0930 16:51:20.734667 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:51:25 crc kubenswrapper[4796]: I0930 16:51:25.848684 4796 generic.go:334] "Generic (PLEG): container finished" podID="0f40019c-f145-46f4-89f3-cc7244d69faf" containerID="b096567650da35b005e4f38d720af409f20acf1c5c5b344ba4f26a3c1b4ea5e9" exitCode=0 Sep 30 16:51:25 crc kubenswrapper[4796]: I0930 16:51:25.848819 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" event={"ID":"0f40019c-f145-46f4-89f3-cc7244d69faf","Type":"ContainerDied","Data":"b096567650da35b005e4f38d720af409f20acf1c5c5b344ba4f26a3c1b4ea5e9"} Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.287617 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.433856 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4lvw\" (UniqueName: \"kubernetes.io/projected/0f40019c-f145-46f4-89f3-cc7244d69faf-kube-api-access-b4lvw\") pod \"0f40019c-f145-46f4-89f3-cc7244d69faf\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.433928 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-inventory\") pod \"0f40019c-f145-46f4-89f3-cc7244d69faf\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.434299 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-ssh-key\") pod \"0f40019c-f145-46f4-89f3-cc7244d69faf\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.434397 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-secret-0\") pod \"0f40019c-f145-46f4-89f3-cc7244d69faf\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.434487 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-combined-ca-bundle\") pod \"0f40019c-f145-46f4-89f3-cc7244d69faf\" (UID: \"0f40019c-f145-46f4-89f3-cc7244d69faf\") " Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.440366 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0f40019c-f145-46f4-89f3-cc7244d69faf" (UID: "0f40019c-f145-46f4-89f3-cc7244d69faf"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.443203 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f40019c-f145-46f4-89f3-cc7244d69faf-kube-api-access-b4lvw" (OuterVolumeSpecName: "kube-api-access-b4lvw") pod "0f40019c-f145-46f4-89f3-cc7244d69faf" (UID: "0f40019c-f145-46f4-89f3-cc7244d69faf"). InnerVolumeSpecName "kube-api-access-b4lvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.467441 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f40019c-f145-46f4-89f3-cc7244d69faf" (UID: "0f40019c-f145-46f4-89f3-cc7244d69faf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.469177 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-inventory" (OuterVolumeSpecName: "inventory") pod "0f40019c-f145-46f4-89f3-cc7244d69faf" (UID: "0f40019c-f145-46f4-89f3-cc7244d69faf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.477423 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "0f40019c-f145-46f4-89f3-cc7244d69faf" (UID: "0f40019c-f145-46f4-89f3-cc7244d69faf"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.537299 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.537356 4796 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.537380 4796 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.537402 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4lvw\" (UniqueName: \"kubernetes.io/projected/0f40019c-f145-46f4-89f3-cc7244d69faf-kube-api-access-b4lvw\") on node \"crc\" DevicePath \"\"" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.537423 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f40019c-f145-46f4-89f3-cc7244d69faf-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.874158 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" event={"ID":"0f40019c-f145-46f4-89f3-cc7244d69faf","Type":"ContainerDied","Data":"2f04d29b4e5e606156ec0d6e0b2a07e3461841e1c789ff6d27924a46815f5973"} Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.874198 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-89pdx" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.874215 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f04d29b4e5e606156ec0d6e0b2a07e3461841e1c789ff6d27924a46815f5973" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.997297 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98"] Sep 30 16:51:27 crc kubenswrapper[4796]: E0930 16:51:27.997910 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="registry-server" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.998066 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="registry-server" Sep 30 16:51:27 crc kubenswrapper[4796]: E0930 16:51:27.998166 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f40019c-f145-46f4-89f3-cc7244d69faf" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.998241 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f40019c-f145-46f4-89f3-cc7244d69faf" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 16:51:27 crc kubenswrapper[4796]: E0930 16:51:27.998341 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="extract-utilities" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.998417 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="extract-utilities" Sep 30 16:51:27 crc kubenswrapper[4796]: E0930 16:51:27.998514 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="extract-content" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.998686 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="extract-content" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.999366 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f40019c-f145-46f4-89f3-cc7244d69faf" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 16:51:27 crc kubenswrapper[4796]: I0930 16:51:27.999470 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="96334f92-e1e6-4261-bc5d-b2310640a925" containerName="registry-server" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.000282 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.002441 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.002481 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.003169 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.003364 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.003706 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.004309 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.004440 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.010494 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98"] Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153106 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153177 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153200 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153230 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153248 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153343 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153366 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r7qh\" (UniqueName: \"kubernetes.io/projected/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-kube-api-access-8r7qh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.153411 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255264 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255423 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255485 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255570 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255614 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255729 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255792 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r7qh\" (UniqueName: \"kubernetes.io/projected/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-kube-api-access-8r7qh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.255935 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.257618 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.261428 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.261648 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.261804 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.263949 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.264183 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.265110 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.271818 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.278904 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r7qh\" (UniqueName: \"kubernetes.io/projected/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-kube-api-access-8r7qh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6gb98\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.315078 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.889850 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98"] Sep 30 16:51:28 crc kubenswrapper[4796]: I0930 16:51:28.897326 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:51:29 crc kubenswrapper[4796]: I0930 16:51:29.894280 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" event={"ID":"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3","Type":"ContainerStarted","Data":"c05675d15a0198afbd5c1d0416197dfc59c471b2b2796eec8845ff2786f1cffa"} Sep 30 16:51:29 crc kubenswrapper[4796]: I0930 16:51:29.894558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" event={"ID":"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3","Type":"ContainerStarted","Data":"0f1a152afee381a3293a84fe2187003b95563a6c10460228c4945349f36fcd69"} Sep 30 16:51:29 crc kubenswrapper[4796]: I0930 16:51:29.914356 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" podStartSLOduration=2.289808736 podStartE2EDuration="2.914341951s" podCreationTimestamp="2025-09-30 16:51:27 +0000 UTC" firstStartedPulling="2025-09-30 16:51:28.897082387 +0000 UTC m=+2380.910360914" lastFinishedPulling="2025-09-30 16:51:29.521615562 +0000 UTC m=+2381.534894129" observedRunningTime="2025-09-30 16:51:29.911730296 +0000 UTC m=+2381.925008823" watchObservedRunningTime="2025-09-30 16:51:29.914341951 +0000 UTC m=+2381.927620478" Sep 30 16:51:31 crc kubenswrapper[4796]: I0930 16:51:31.733069 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:51:31 crc kubenswrapper[4796]: E0930 16:51:31.734087 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:51:43 crc kubenswrapper[4796]: I0930 16:51:43.734412 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:51:43 crc kubenswrapper[4796]: E0930 16:51:43.735742 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:51:54 crc kubenswrapper[4796]: I0930 16:51:54.733736 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:51:54 crc kubenswrapper[4796]: E0930 16:51:54.734921 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:52:08 crc kubenswrapper[4796]: I0930 16:52:08.746471 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:52:08 crc kubenswrapper[4796]: E0930 16:52:08.747773 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:52:23 crc kubenswrapper[4796]: I0930 16:52:23.734166 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:52:23 crc kubenswrapper[4796]: E0930 16:52:23.735036 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:52:36 crc kubenswrapper[4796]: I0930 16:52:36.733873 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:52:36 crc kubenswrapper[4796]: E0930 16:52:36.734690 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:52:49 crc kubenswrapper[4796]: I0930 16:52:49.733214 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:52:49 crc kubenswrapper[4796]: E0930 16:52:49.733938 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:53:01 crc kubenswrapper[4796]: I0930 16:53:01.733281 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:53:02 crc kubenswrapper[4796]: I0930 16:53:02.901438 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"aa54084bf09f071e4795b1854e506cfb9ca943c018b90c5bd451399efd8debd9"} Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.496135 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vzhbz"] Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.499107 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.512354 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzhbz"] Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.535202 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-utilities\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.535343 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-catalog-content\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.535503 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fmwt\" (UniqueName: \"kubernetes.io/projected/314ccb80-a820-4392-adf9-e01a50c9441e-kube-api-access-8fmwt\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.637635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-catalog-content\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.637708 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fmwt\" (UniqueName: \"kubernetes.io/projected/314ccb80-a820-4392-adf9-e01a50c9441e-kube-api-access-8fmwt\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.637798 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-utilities\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.638446 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-catalog-content\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.638783 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-utilities\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.663001 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fmwt\" (UniqueName: \"kubernetes.io/projected/314ccb80-a820-4392-adf9-e01a50c9441e-kube-api-access-8fmwt\") pod \"certified-operators-vzhbz\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:02 crc kubenswrapper[4796]: I0930 16:54:02.822882 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:03 crc kubenswrapper[4796]: I0930 16:54:03.370175 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzhbz"] Sep 30 16:54:03 crc kubenswrapper[4796]: I0930 16:54:03.576918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerStarted","Data":"2210b5c7ce77a801819946c210492e3e72369390efa524509668cd38ae45489d"} Sep 30 16:54:04 crc kubenswrapper[4796]: I0930 16:54:04.587118 4796 generic.go:334] "Generic (PLEG): container finished" podID="314ccb80-a820-4392-adf9-e01a50c9441e" containerID="7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9" exitCode=0 Sep 30 16:54:04 crc kubenswrapper[4796]: I0930 16:54:04.587180 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerDied","Data":"7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9"} Sep 30 16:54:05 crc kubenswrapper[4796]: I0930 16:54:05.597938 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerStarted","Data":"eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b"} Sep 30 16:54:06 crc kubenswrapper[4796]: I0930 16:54:06.614515 4796 generic.go:334] "Generic (PLEG): container finished" podID="314ccb80-a820-4392-adf9-e01a50c9441e" containerID="eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b" exitCode=0 Sep 30 16:54:06 crc kubenswrapper[4796]: I0930 16:54:06.614573 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerDied","Data":"eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b"} Sep 30 16:54:07 crc kubenswrapper[4796]: I0930 16:54:07.626726 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerStarted","Data":"db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17"} Sep 30 16:54:07 crc kubenswrapper[4796]: I0930 16:54:07.661741 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vzhbz" podStartSLOduration=3.214390707 podStartE2EDuration="5.661710216s" podCreationTimestamp="2025-09-30 16:54:02 +0000 UTC" firstStartedPulling="2025-09-30 16:54:04.590104475 +0000 UTC m=+2536.603383032" lastFinishedPulling="2025-09-30 16:54:07.037424014 +0000 UTC m=+2539.050702541" observedRunningTime="2025-09-30 16:54:07.651412709 +0000 UTC m=+2539.664691256" watchObservedRunningTime="2025-09-30 16:54:07.661710216 +0000 UTC m=+2539.674988783" Sep 30 16:54:12 crc kubenswrapper[4796]: I0930 16:54:12.823596 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:12 crc kubenswrapper[4796]: I0930 16:54:12.824356 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:12 crc kubenswrapper[4796]: I0930 16:54:12.913253 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:13 crc kubenswrapper[4796]: I0930 16:54:13.781442 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:14 crc kubenswrapper[4796]: I0930 16:54:14.277717 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzhbz"] Sep 30 16:54:15 crc kubenswrapper[4796]: I0930 16:54:15.717430 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vzhbz" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="registry-server" containerID="cri-o://db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17" gracePeriod=2 Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.193952 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.243910 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-utilities\") pod \"314ccb80-a820-4392-adf9-e01a50c9441e\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.244256 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-catalog-content\") pod \"314ccb80-a820-4392-adf9-e01a50c9441e\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.244385 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fmwt\" (UniqueName: \"kubernetes.io/projected/314ccb80-a820-4392-adf9-e01a50c9441e-kube-api-access-8fmwt\") pod \"314ccb80-a820-4392-adf9-e01a50c9441e\" (UID: \"314ccb80-a820-4392-adf9-e01a50c9441e\") " Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.248934 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-utilities" (OuterVolumeSpecName: "utilities") pod "314ccb80-a820-4392-adf9-e01a50c9441e" (UID: "314ccb80-a820-4392-adf9-e01a50c9441e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.255733 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314ccb80-a820-4392-adf9-e01a50c9441e-kube-api-access-8fmwt" (OuterVolumeSpecName: "kube-api-access-8fmwt") pod "314ccb80-a820-4392-adf9-e01a50c9441e" (UID: "314ccb80-a820-4392-adf9-e01a50c9441e"). InnerVolumeSpecName "kube-api-access-8fmwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.346460 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.346496 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fmwt\" (UniqueName: \"kubernetes.io/projected/314ccb80-a820-4392-adf9-e01a50c9441e-kube-api-access-8fmwt\") on node \"crc\" DevicePath \"\"" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.381713 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "314ccb80-a820-4392-adf9-e01a50c9441e" (UID: "314ccb80-a820-4392-adf9-e01a50c9441e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.448251 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314ccb80-a820-4392-adf9-e01a50c9441e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.729471 4796 generic.go:334] "Generic (PLEG): container finished" podID="314ccb80-a820-4392-adf9-e01a50c9441e" containerID="db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17" exitCode=0 Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.729516 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzhbz" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.729570 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerDied","Data":"db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17"} Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.729909 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzhbz" event={"ID":"314ccb80-a820-4392-adf9-e01a50c9441e","Type":"ContainerDied","Data":"2210b5c7ce77a801819946c210492e3e72369390efa524509668cd38ae45489d"} Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.729940 4796 scope.go:117] "RemoveContainer" containerID="db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.767653 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzhbz"] Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.768425 4796 scope.go:117] "RemoveContainer" containerID="eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.775906 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vzhbz"] Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.806516 4796 scope.go:117] "RemoveContainer" containerID="7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.851546 4796 scope.go:117] "RemoveContainer" containerID="db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17" Sep 30 16:54:16 crc kubenswrapper[4796]: E0930 16:54:16.852124 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17\": container with ID starting with db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17 not found: ID does not exist" containerID="db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.852209 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17"} err="failed to get container status \"db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17\": rpc error: code = NotFound desc = could not find container \"db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17\": container with ID starting with db6ff157d4ef72dda241f99966e9255c7a630a0834045763c901b66644525a17 not found: ID does not exist" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.852247 4796 scope.go:117] "RemoveContainer" containerID="eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b" Sep 30 16:54:16 crc kubenswrapper[4796]: E0930 16:54:16.852738 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b\": container with ID starting with eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b not found: ID does not exist" containerID="eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.852770 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b"} err="failed to get container status \"eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b\": rpc error: code = NotFound desc = could not find container \"eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b\": container with ID starting with eee9e96632114767d897dcc7305c6553278a9a2af0e4ebd128986084e8bf928b not found: ID does not exist" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.852790 4796 scope.go:117] "RemoveContainer" containerID="7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9" Sep 30 16:54:16 crc kubenswrapper[4796]: E0930 16:54:16.853184 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9\": container with ID starting with 7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9 not found: ID does not exist" containerID="7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9" Sep 30 16:54:16 crc kubenswrapper[4796]: I0930 16:54:16.853213 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9"} err="failed to get container status \"7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9\": rpc error: code = NotFound desc = could not find container \"7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9\": container with ID starting with 7233d0d4df953a1d9551fb803f965d12ed45559f17415378b5685707465a34c9 not found: ID does not exist" Sep 30 16:54:18 crc kubenswrapper[4796]: I0930 16:54:18.746552 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" path="/var/lib/kubelet/pods/314ccb80-a820-4392-adf9-e01a50c9441e/volumes" Sep 30 16:55:01 crc kubenswrapper[4796]: I0930 16:55:01.097602 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:55:01 crc kubenswrapper[4796]: I0930 16:55:01.098240 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:55:09 crc kubenswrapper[4796]: I0930 16:55:09.264315 4796 generic.go:334] "Generic (PLEG): container finished" podID="6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" containerID="c05675d15a0198afbd5c1d0416197dfc59c471b2b2796eec8845ff2786f1cffa" exitCode=0 Sep 30 16:55:09 crc kubenswrapper[4796]: I0930 16:55:09.264400 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" event={"ID":"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3","Type":"ContainerDied","Data":"c05675d15a0198afbd5c1d0416197dfc59c471b2b2796eec8845ff2786f1cffa"} Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.729014 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769417 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-1\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769483 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-0\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769577 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-extra-config-0\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769649 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r7qh\" (UniqueName: \"kubernetes.io/projected/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-kube-api-access-8r7qh\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769690 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-0\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769714 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-ssh-key\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769843 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-1\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.769904 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-combined-ca-bundle\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.770001 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-inventory\") pod \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\" (UID: \"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3\") " Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.784186 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.786151 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-kube-api-access-8r7qh" (OuterVolumeSpecName: "kube-api-access-8r7qh") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "kube-api-access-8r7qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.800300 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.814869 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.817673 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.827150 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.830261 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.834382 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.849040 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-inventory" (OuterVolumeSpecName: "inventory") pod "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" (UID: "6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872601 4796 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872650 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872666 4796 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872679 4796 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872691 4796 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872705 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r7qh\" (UniqueName: \"kubernetes.io/projected/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-kube-api-access-8r7qh\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872718 4796 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872731 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:10 crc kubenswrapper[4796]: I0930 16:55:10.872743 4796 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.285016 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" event={"ID":"6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3","Type":"ContainerDied","Data":"0f1a152afee381a3293a84fe2187003b95563a6c10460228c4945349f36fcd69"} Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.285069 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f1a152afee381a3293a84fe2187003b95563a6c10460228c4945349f36fcd69" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.285135 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6gb98" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.417410 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p"] Sep 30 16:55:11 crc kubenswrapper[4796]: E0930 16:55:11.418270 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="extract-utilities" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.418375 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="extract-utilities" Sep 30 16:55:11 crc kubenswrapper[4796]: E0930 16:55:11.418431 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.418483 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 16:55:11 crc kubenswrapper[4796]: E0930 16:55:11.418568 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="registry-server" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.418634 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="registry-server" Sep 30 16:55:11 crc kubenswrapper[4796]: E0930 16:55:11.418708 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="extract-content" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.418779 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="extract-content" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.419139 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="314ccb80-a820-4392-adf9-e01a50c9441e" containerName="registry-server" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.419218 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.419847 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.423368 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.423576 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vpch6" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.423766 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.423928 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.424393 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.433814 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p"] Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.484877 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.484946 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pd75\" (UniqueName: \"kubernetes.io/projected/e8acd478-67ee-465a-b801-6731a62328f3-kube-api-access-2pd75\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.485229 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.485349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.485415 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.485608 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.485646 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.587508 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.587806 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pd75\" (UniqueName: \"kubernetes.io/projected/e8acd478-67ee-465a-b801-6731a62328f3-kube-api-access-2pd75\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.587958 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.588081 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.588179 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.588304 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.588380 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.594785 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.595147 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.596145 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.596779 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.598214 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.598886 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.612187 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pd75\" (UniqueName: \"kubernetes.io/projected/e8acd478-67ee-465a-b801-6731a62328f3-kube-api-access-2pd75\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:11 crc kubenswrapper[4796]: I0930 16:55:11.739996 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:55:12 crc kubenswrapper[4796]: I0930 16:55:12.441829 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p"] Sep 30 16:55:13 crc kubenswrapper[4796]: I0930 16:55:13.304413 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" event={"ID":"e8acd478-67ee-465a-b801-6731a62328f3","Type":"ContainerStarted","Data":"103dbeadb4570aa925d439ed98844e3042251360153db6a43549e95fdff6feea"} Sep 30 16:55:14 crc kubenswrapper[4796]: I0930 16:55:14.318906 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" event={"ID":"e8acd478-67ee-465a-b801-6731a62328f3","Type":"ContainerStarted","Data":"cf53529e111ef1032f81542560dbd2b09f8462543e106d2a5371bdd7fd4bfe1c"} Sep 30 16:55:14 crc kubenswrapper[4796]: I0930 16:55:14.352573 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" podStartSLOduration=2.502738315 podStartE2EDuration="3.352550384s" podCreationTimestamp="2025-09-30 16:55:11 +0000 UTC" firstStartedPulling="2025-09-30 16:55:12.455890245 +0000 UTC m=+2604.469168772" lastFinishedPulling="2025-09-30 16:55:13.305702304 +0000 UTC m=+2605.318980841" observedRunningTime="2025-09-30 16:55:14.33627002 +0000 UTC m=+2606.349548537" watchObservedRunningTime="2025-09-30 16:55:14.352550384 +0000 UTC m=+2606.365828921" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.188715 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m4wdc"] Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.192297 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.226171 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4wdc"] Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.226381 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-catalog-content\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.226645 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-utilities\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.226693 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzz87\" (UniqueName: \"kubernetes.io/projected/f4a4bc1f-3935-4978-b72b-2ef962beb284-kube-api-access-dzz87\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.329254 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-catalog-content\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.329375 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-utilities\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.329401 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzz87\" (UniqueName: \"kubernetes.io/projected/f4a4bc1f-3935-4978-b72b-2ef962beb284-kube-api-access-dzz87\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.329686 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-catalog-content\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.329766 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-utilities\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.355897 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzz87\" (UniqueName: \"kubernetes.io/projected/f4a4bc1f-3935-4978-b72b-2ef962beb284-kube-api-access-dzz87\") pod \"community-operators-m4wdc\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:28 crc kubenswrapper[4796]: I0930 16:55:28.536153 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:29 crc kubenswrapper[4796]: I0930 16:55:29.026712 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4wdc"] Sep 30 16:55:29 crc kubenswrapper[4796]: I0930 16:55:29.473699 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerID="61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1" exitCode=0 Sep 30 16:55:29 crc kubenswrapper[4796]: I0930 16:55:29.473806 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerDied","Data":"61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1"} Sep 30 16:55:29 crc kubenswrapper[4796]: I0930 16:55:29.474088 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerStarted","Data":"e4073f49523aa5cc149d265eb6d3b12e26e5d827a3d73692e688b0b9fdfe3849"} Sep 30 16:55:30 crc kubenswrapper[4796]: I0930 16:55:30.483369 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerStarted","Data":"20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6"} Sep 30 16:55:31 crc kubenswrapper[4796]: I0930 16:55:31.096731 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:55:31 crc kubenswrapper[4796]: I0930 16:55:31.096858 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:55:31 crc kubenswrapper[4796]: I0930 16:55:31.503322 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerID="20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6" exitCode=0 Sep 30 16:55:31 crc kubenswrapper[4796]: I0930 16:55:31.503378 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerDied","Data":"20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6"} Sep 30 16:55:32 crc kubenswrapper[4796]: I0930 16:55:32.515724 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerStarted","Data":"6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3"} Sep 30 16:55:32 crc kubenswrapper[4796]: I0930 16:55:32.533964 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m4wdc" podStartSLOduration=1.718360374 podStartE2EDuration="4.533939288s" podCreationTimestamp="2025-09-30 16:55:28 +0000 UTC" firstStartedPulling="2025-09-30 16:55:29.476258555 +0000 UTC m=+2621.489537122" lastFinishedPulling="2025-09-30 16:55:32.291837499 +0000 UTC m=+2624.305116036" observedRunningTime="2025-09-30 16:55:32.530302577 +0000 UTC m=+2624.543581124" watchObservedRunningTime="2025-09-30 16:55:32.533939288 +0000 UTC m=+2624.547217835" Sep 30 16:55:38 crc kubenswrapper[4796]: I0930 16:55:38.536602 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:38 crc kubenswrapper[4796]: I0930 16:55:38.537175 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:38 crc kubenswrapper[4796]: I0930 16:55:38.592346 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:38 crc kubenswrapper[4796]: I0930 16:55:38.636148 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:38 crc kubenswrapper[4796]: I0930 16:55:38.832915 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m4wdc"] Sep 30 16:55:40 crc kubenswrapper[4796]: I0930 16:55:40.606017 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m4wdc" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="registry-server" containerID="cri-o://6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3" gracePeriod=2 Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.097951 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.189893 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-catalog-content\") pod \"f4a4bc1f-3935-4978-b72b-2ef962beb284\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.190168 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-utilities\") pod \"f4a4bc1f-3935-4978-b72b-2ef962beb284\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.190269 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzz87\" (UniqueName: \"kubernetes.io/projected/f4a4bc1f-3935-4978-b72b-2ef962beb284-kube-api-access-dzz87\") pod \"f4a4bc1f-3935-4978-b72b-2ef962beb284\" (UID: \"f4a4bc1f-3935-4978-b72b-2ef962beb284\") " Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.191413 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-utilities" (OuterVolumeSpecName: "utilities") pod "f4a4bc1f-3935-4978-b72b-2ef962beb284" (UID: "f4a4bc1f-3935-4978-b72b-2ef962beb284"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.196919 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a4bc1f-3935-4978-b72b-2ef962beb284-kube-api-access-dzz87" (OuterVolumeSpecName: "kube-api-access-dzz87") pod "f4a4bc1f-3935-4978-b72b-2ef962beb284" (UID: "f4a4bc1f-3935-4978-b72b-2ef962beb284"). InnerVolumeSpecName "kube-api-access-dzz87". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.293626 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.293679 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzz87\" (UniqueName: \"kubernetes.io/projected/f4a4bc1f-3935-4978-b72b-2ef962beb284-kube-api-access-dzz87\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.620376 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerID="6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3" exitCode=0 Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.620471 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4wdc" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.620442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerDied","Data":"6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3"} Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.620884 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4wdc" event={"ID":"f4a4bc1f-3935-4978-b72b-2ef962beb284","Type":"ContainerDied","Data":"e4073f49523aa5cc149d265eb6d3b12e26e5d827a3d73692e688b0b9fdfe3849"} Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.620956 4796 scope.go:117] "RemoveContainer" containerID="6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.642344 4796 scope.go:117] "RemoveContainer" containerID="20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.677390 4796 scope.go:117] "RemoveContainer" containerID="61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.723468 4796 scope.go:117] "RemoveContainer" containerID="6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3" Sep 30 16:55:41 crc kubenswrapper[4796]: E0930 16:55:41.723889 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3\": container with ID starting with 6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3 not found: ID does not exist" containerID="6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.723949 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3"} err="failed to get container status \"6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3\": rpc error: code = NotFound desc = could not find container \"6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3\": container with ID starting with 6d7e605555de48f0d443fec59a0af4f60658255a9494943dbd4ec1f5ebd5f4b3 not found: ID does not exist" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.723997 4796 scope.go:117] "RemoveContainer" containerID="20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6" Sep 30 16:55:41 crc kubenswrapper[4796]: E0930 16:55:41.724469 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6\": container with ID starting with 20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6 not found: ID does not exist" containerID="20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.724503 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6"} err="failed to get container status \"20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6\": rpc error: code = NotFound desc = could not find container \"20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6\": container with ID starting with 20eec181d4d929ca260b9f077db49525b3073a2672d855c9d431a1c764b48ac6 not found: ID does not exist" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.724523 4796 scope.go:117] "RemoveContainer" containerID="61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1" Sep 30 16:55:41 crc kubenswrapper[4796]: E0930 16:55:41.724806 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1\": container with ID starting with 61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1 not found: ID does not exist" containerID="61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1" Sep 30 16:55:41 crc kubenswrapper[4796]: I0930 16:55:41.724846 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1"} err="failed to get container status \"61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1\": rpc error: code = NotFound desc = could not find container \"61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1\": container with ID starting with 61dd7c1bac2378bc8828dbc802cb14bb4be789c1df04fee4b9b38fd91d17fef1 not found: ID does not exist" Sep 30 16:55:42 crc kubenswrapper[4796]: I0930 16:55:42.017203 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4a4bc1f-3935-4978-b72b-2ef962beb284" (UID: "f4a4bc1f-3935-4978-b72b-2ef962beb284"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:55:42 crc kubenswrapper[4796]: I0930 16:55:42.106248 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a4bc1f-3935-4978-b72b-2ef962beb284-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:55:42 crc kubenswrapper[4796]: I0930 16:55:42.278545 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m4wdc"] Sep 30 16:55:42 crc kubenswrapper[4796]: I0930 16:55:42.287693 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m4wdc"] Sep 30 16:55:42 crc kubenswrapper[4796]: I0930 16:55:42.747545 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" path="/var/lib/kubelet/pods/f4a4bc1f-3935-4978-b72b-2ef962beb284/volumes" Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.097559 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.098174 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.098229 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.099123 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa54084bf09f071e4795b1854e506cfb9ca943c018b90c5bd451399efd8debd9"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.099190 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://aa54084bf09f071e4795b1854e506cfb9ca943c018b90c5bd451399efd8debd9" gracePeriod=600 Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.827320 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="aa54084bf09f071e4795b1854e506cfb9ca943c018b90c5bd451399efd8debd9" exitCode=0 Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.827390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"aa54084bf09f071e4795b1854e506cfb9ca943c018b90c5bd451399efd8debd9"} Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.827939 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708"} Sep 30 16:56:01 crc kubenswrapper[4796]: I0930 16:56:01.827969 4796 scope.go:117] "RemoveContainer" containerID="3d776b21cc6ac8065373a16433f445daf92282ab7b6979216211698203cc7752" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.914229 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9w6v8"] Sep 30 16:56:23 crc kubenswrapper[4796]: E0930 16:56:23.915281 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="extract-content" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.915299 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="extract-content" Sep 30 16:56:23 crc kubenswrapper[4796]: E0930 16:56:23.915331 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="extract-utilities" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.915340 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="extract-utilities" Sep 30 16:56:23 crc kubenswrapper[4796]: E0930 16:56:23.915363 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="registry-server" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.915372 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="registry-server" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.915623 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a4bc1f-3935-4978-b72b-2ef962beb284" containerName="registry-server" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.917929 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.986950 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-catalog-content\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.987564 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-utilities\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.987600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2r8f\" (UniqueName: \"kubernetes.io/projected/53fee086-770e-436a-93e5-5e3b05f0658c-kube-api-access-x2r8f\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:23 crc kubenswrapper[4796]: I0930 16:56:23.993903 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9w6v8"] Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.089187 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-utilities\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.089240 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2r8f\" (UniqueName: \"kubernetes.io/projected/53fee086-770e-436a-93e5-5e3b05f0658c-kube-api-access-x2r8f\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.089284 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-catalog-content\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.089911 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-catalog-content\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.090653 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-utilities\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.108556 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2r8f\" (UniqueName: \"kubernetes.io/projected/53fee086-770e-436a-93e5-5e3b05f0658c-kube-api-access-x2r8f\") pod \"redhat-operators-9w6v8\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.294692 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:24 crc kubenswrapper[4796]: I0930 16:56:24.762285 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9w6v8"] Sep 30 16:56:25 crc kubenswrapper[4796]: I0930 16:56:25.054179 4796 generic.go:334] "Generic (PLEG): container finished" podID="53fee086-770e-436a-93e5-5e3b05f0658c" containerID="a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a" exitCode=0 Sep 30 16:56:25 crc kubenswrapper[4796]: I0930 16:56:25.054494 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9w6v8" event={"ID":"53fee086-770e-436a-93e5-5e3b05f0658c","Type":"ContainerDied","Data":"a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a"} Sep 30 16:56:25 crc kubenswrapper[4796]: I0930 16:56:25.054525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9w6v8" event={"ID":"53fee086-770e-436a-93e5-5e3b05f0658c","Type":"ContainerStarted","Data":"42a01442511b7ee73a858e31aabc6955200bba5d5643358c793a75f2fcb5cb15"} Sep 30 16:56:27 crc kubenswrapper[4796]: I0930 16:56:27.074349 4796 generic.go:334] "Generic (PLEG): container finished" podID="53fee086-770e-436a-93e5-5e3b05f0658c" containerID="56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6" exitCode=0 Sep 30 16:56:27 crc kubenswrapper[4796]: I0930 16:56:27.074443 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9w6v8" event={"ID":"53fee086-770e-436a-93e5-5e3b05f0658c","Type":"ContainerDied","Data":"56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6"} Sep 30 16:56:28 crc kubenswrapper[4796]: I0930 16:56:28.087962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9w6v8" event={"ID":"53fee086-770e-436a-93e5-5e3b05f0658c","Type":"ContainerStarted","Data":"51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9"} Sep 30 16:56:28 crc kubenswrapper[4796]: I0930 16:56:28.118342 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9w6v8" podStartSLOduration=2.565796974 podStartE2EDuration="5.118322096s" podCreationTimestamp="2025-09-30 16:56:23 +0000 UTC" firstStartedPulling="2025-09-30 16:56:25.056918599 +0000 UTC m=+2677.070197126" lastFinishedPulling="2025-09-30 16:56:27.609443721 +0000 UTC m=+2679.622722248" observedRunningTime="2025-09-30 16:56:28.109195451 +0000 UTC m=+2680.122473978" watchObservedRunningTime="2025-09-30 16:56:28.118322096 +0000 UTC m=+2680.131600623" Sep 30 16:56:34 crc kubenswrapper[4796]: I0930 16:56:34.295015 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:34 crc kubenswrapper[4796]: I0930 16:56:34.295448 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:34 crc kubenswrapper[4796]: I0930 16:56:34.355048 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:35 crc kubenswrapper[4796]: I0930 16:56:35.216653 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:35 crc kubenswrapper[4796]: I0930 16:56:35.264203 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9w6v8"] Sep 30 16:56:37 crc kubenswrapper[4796]: I0930 16:56:37.171017 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9w6v8" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="registry-server" containerID="cri-o://51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9" gracePeriod=2 Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.100742 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.183322 4796 generic.go:334] "Generic (PLEG): container finished" podID="53fee086-770e-436a-93e5-5e3b05f0658c" containerID="51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9" exitCode=0 Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.183362 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9w6v8" event={"ID":"53fee086-770e-436a-93e5-5e3b05f0658c","Type":"ContainerDied","Data":"51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9"} Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.183387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9w6v8" event={"ID":"53fee086-770e-436a-93e5-5e3b05f0658c","Type":"ContainerDied","Data":"42a01442511b7ee73a858e31aabc6955200bba5d5643358c793a75f2fcb5cb15"} Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.183407 4796 scope.go:117] "RemoveContainer" containerID="51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.184628 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9w6v8" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.197802 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2r8f\" (UniqueName: \"kubernetes.io/projected/53fee086-770e-436a-93e5-5e3b05f0658c-kube-api-access-x2r8f\") pod \"53fee086-770e-436a-93e5-5e3b05f0658c\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.198149 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-catalog-content\") pod \"53fee086-770e-436a-93e5-5e3b05f0658c\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.198223 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-utilities\") pod \"53fee086-770e-436a-93e5-5e3b05f0658c\" (UID: \"53fee086-770e-436a-93e5-5e3b05f0658c\") " Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.199000 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-utilities" (OuterVolumeSpecName: "utilities") pod "53fee086-770e-436a-93e5-5e3b05f0658c" (UID: "53fee086-770e-436a-93e5-5e3b05f0658c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.206882 4796 scope.go:117] "RemoveContainer" containerID="56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.207523 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53fee086-770e-436a-93e5-5e3b05f0658c-kube-api-access-x2r8f" (OuterVolumeSpecName: "kube-api-access-x2r8f") pod "53fee086-770e-436a-93e5-5e3b05f0658c" (UID: "53fee086-770e-436a-93e5-5e3b05f0658c"). InnerVolumeSpecName "kube-api-access-x2r8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.266595 4796 scope.go:117] "RemoveContainer" containerID="a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.286763 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53fee086-770e-436a-93e5-5e3b05f0658c" (UID: "53fee086-770e-436a-93e5-5e3b05f0658c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.300851 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.300889 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53fee086-770e-436a-93e5-5e3b05f0658c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.300904 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2r8f\" (UniqueName: \"kubernetes.io/projected/53fee086-770e-436a-93e5-5e3b05f0658c-kube-api-access-x2r8f\") on node \"crc\" DevicePath \"\"" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.308202 4796 scope.go:117] "RemoveContainer" containerID="51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9" Sep 30 16:56:38 crc kubenswrapper[4796]: E0930 16:56:38.308806 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9\": container with ID starting with 51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9 not found: ID does not exist" containerID="51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.308899 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9"} err="failed to get container status \"51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9\": rpc error: code = NotFound desc = could not find container \"51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9\": container with ID starting with 51d9c8d2a9c5ce5c8ba263a2ac83a6b194f5fb6b9ddd0116e39e266061b70de9 not found: ID does not exist" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.308928 4796 scope.go:117] "RemoveContainer" containerID="56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6" Sep 30 16:56:38 crc kubenswrapper[4796]: E0930 16:56:38.309268 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6\": container with ID starting with 56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6 not found: ID does not exist" containerID="56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.309302 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6"} err="failed to get container status \"56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6\": rpc error: code = NotFound desc = could not find container \"56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6\": container with ID starting with 56cbae0e54be9687b207540ebe6f5d318156592b225349d7dcc6d10c33bdd6e6 not found: ID does not exist" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.309324 4796 scope.go:117] "RemoveContainer" containerID="a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a" Sep 30 16:56:38 crc kubenswrapper[4796]: E0930 16:56:38.309603 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a\": container with ID starting with a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a not found: ID does not exist" containerID="a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.309626 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a"} err="failed to get container status \"a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a\": rpc error: code = NotFound desc = could not find container \"a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a\": container with ID starting with a2df8c577da87852ec36c25a7699c16bc78f5993c5cc7b75069f7bd2c67cbe0a not found: ID does not exist" Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.535213 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9w6v8"] Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.545472 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9w6v8"] Sep 30 16:56:38 crc kubenswrapper[4796]: I0930 16:56:38.750407 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" path="/var/lib/kubelet/pods/53fee086-770e-436a-93e5-5e3b05f0658c/volumes" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.670136 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qdfzp"] Sep 30 16:57:27 crc kubenswrapper[4796]: E0930 16:57:27.671128 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="registry-server" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.671149 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="registry-server" Sep 30 16:57:27 crc kubenswrapper[4796]: E0930 16:57:27.671164 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="extract-content" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.671172 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="extract-content" Sep 30 16:57:27 crc kubenswrapper[4796]: E0930 16:57:27.671185 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="extract-utilities" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.671192 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="extract-utilities" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.671450 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="53fee086-770e-436a-93e5-5e3b05f0658c" containerName="registry-server" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.676024 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.696441 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdfzp"] Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.759237 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-catalog-content\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.759335 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-utilities\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.759504 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmz4k\" (UniqueName: \"kubernetes.io/projected/58c16fe7-0b61-44f3-b089-4195852e1a09-kube-api-access-zmz4k\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.861116 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-utilities\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.861282 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmz4k\" (UniqueName: \"kubernetes.io/projected/58c16fe7-0b61-44f3-b089-4195852e1a09-kube-api-access-zmz4k\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.861529 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-catalog-content\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.861656 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-utilities\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.862346 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-catalog-content\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.885962 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmz4k\" (UniqueName: \"kubernetes.io/projected/58c16fe7-0b61-44f3-b089-4195852e1a09-kube-api-access-zmz4k\") pod \"redhat-marketplace-qdfzp\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:27 crc kubenswrapper[4796]: I0930 16:57:27.995491 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:28 crc kubenswrapper[4796]: I0930 16:57:28.449385 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdfzp"] Sep 30 16:57:28 crc kubenswrapper[4796]: I0930 16:57:28.715430 4796 generic.go:334] "Generic (PLEG): container finished" podID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerID="36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6" exitCode=0 Sep 30 16:57:28 crc kubenswrapper[4796]: I0930 16:57:28.715481 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerDied","Data":"36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6"} Sep 30 16:57:28 crc kubenswrapper[4796]: I0930 16:57:28.715511 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerStarted","Data":"5bbe8e9e03513c63575be8d08c6512b5fe08ce858d62b1bfe458352ff1dbb605"} Sep 30 16:57:28 crc kubenswrapper[4796]: I0930 16:57:28.717829 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 16:57:29 crc kubenswrapper[4796]: I0930 16:57:29.725448 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerStarted","Data":"f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4"} Sep 30 16:57:30 crc kubenswrapper[4796]: I0930 16:57:30.743967 4796 generic.go:334] "Generic (PLEG): container finished" podID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerID="f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4" exitCode=0 Sep 30 16:57:30 crc kubenswrapper[4796]: I0930 16:57:30.745714 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerDied","Data":"f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4"} Sep 30 16:57:31 crc kubenswrapper[4796]: I0930 16:57:31.756209 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerStarted","Data":"d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c"} Sep 30 16:57:31 crc kubenswrapper[4796]: I0930 16:57:31.787623 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qdfzp" podStartSLOduration=2.210281755 podStartE2EDuration="4.787597677s" podCreationTimestamp="2025-09-30 16:57:27 +0000 UTC" firstStartedPulling="2025-09-30 16:57:28.71755246 +0000 UTC m=+2740.730830987" lastFinishedPulling="2025-09-30 16:57:31.294868382 +0000 UTC m=+2743.308146909" observedRunningTime="2025-09-30 16:57:31.780663034 +0000 UTC m=+2743.793941561" watchObservedRunningTime="2025-09-30 16:57:31.787597677 +0000 UTC m=+2743.800876224" Sep 30 16:57:37 crc kubenswrapper[4796]: I0930 16:57:37.996300 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:37 crc kubenswrapper[4796]: I0930 16:57:37.996900 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:38 crc kubenswrapper[4796]: I0930 16:57:38.046123 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:38 crc kubenswrapper[4796]: I0930 16:57:38.879951 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:38 crc kubenswrapper[4796]: I0930 16:57:38.939944 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdfzp"] Sep 30 16:57:40 crc kubenswrapper[4796]: I0930 16:57:40.837718 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qdfzp" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="registry-server" containerID="cri-o://d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c" gracePeriod=2 Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.401901 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.559966 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-utilities\") pod \"58c16fe7-0b61-44f3-b089-4195852e1a09\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.560128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-catalog-content\") pod \"58c16fe7-0b61-44f3-b089-4195852e1a09\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.560197 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmz4k\" (UniqueName: \"kubernetes.io/projected/58c16fe7-0b61-44f3-b089-4195852e1a09-kube-api-access-zmz4k\") pod \"58c16fe7-0b61-44f3-b089-4195852e1a09\" (UID: \"58c16fe7-0b61-44f3-b089-4195852e1a09\") " Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.560819 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-utilities" (OuterVolumeSpecName: "utilities") pod "58c16fe7-0b61-44f3-b089-4195852e1a09" (UID: "58c16fe7-0b61-44f3-b089-4195852e1a09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.570694 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c16fe7-0b61-44f3-b089-4195852e1a09-kube-api-access-zmz4k" (OuterVolumeSpecName: "kube-api-access-zmz4k") pod "58c16fe7-0b61-44f3-b089-4195852e1a09" (UID: "58c16fe7-0b61-44f3-b089-4195852e1a09"). InnerVolumeSpecName "kube-api-access-zmz4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.573274 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58c16fe7-0b61-44f3-b089-4195852e1a09" (UID: "58c16fe7-0b61-44f3-b089-4195852e1a09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.662580 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.662608 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c16fe7-0b61-44f3-b089-4195852e1a09-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.662620 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmz4k\" (UniqueName: \"kubernetes.io/projected/58c16fe7-0b61-44f3-b089-4195852e1a09-kube-api-access-zmz4k\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.848588 4796 generic.go:334] "Generic (PLEG): container finished" podID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerID="d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c" exitCode=0 Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.848634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerDied","Data":"d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c"} Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.848663 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdfzp" event={"ID":"58c16fe7-0b61-44f3-b089-4195852e1a09","Type":"ContainerDied","Data":"5bbe8e9e03513c63575be8d08c6512b5fe08ce858d62b1bfe458352ff1dbb605"} Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.848686 4796 scope.go:117] "RemoveContainer" containerID="d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.848684 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdfzp" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.886713 4796 scope.go:117] "RemoveContainer" containerID="f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.906371 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdfzp"] Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.925306 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdfzp"] Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.931591 4796 scope.go:117] "RemoveContainer" containerID="36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.969200 4796 scope.go:117] "RemoveContainer" containerID="d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c" Sep 30 16:57:41 crc kubenswrapper[4796]: E0930 16:57:41.969546 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c\": container with ID starting with d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c not found: ID does not exist" containerID="d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.969580 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c"} err="failed to get container status \"d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c\": rpc error: code = NotFound desc = could not find container \"d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c\": container with ID starting with d83faa158bd4c62df7b88ff7aafda9ca71edf00fd8ee4924f943e84c791ac74c not found: ID does not exist" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.969609 4796 scope.go:117] "RemoveContainer" containerID="f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4" Sep 30 16:57:41 crc kubenswrapper[4796]: E0930 16:57:41.970473 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4\": container with ID starting with f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4 not found: ID does not exist" containerID="f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.970499 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4"} err="failed to get container status \"f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4\": rpc error: code = NotFound desc = could not find container \"f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4\": container with ID starting with f0a66a7b99c78fe2a87d480f839033d8ab80d2265b33b13b3703e50dbac71db4 not found: ID does not exist" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.970513 4796 scope.go:117] "RemoveContainer" containerID="36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6" Sep 30 16:57:41 crc kubenswrapper[4796]: E0930 16:57:41.970830 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6\": container with ID starting with 36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6 not found: ID does not exist" containerID="36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6" Sep 30 16:57:41 crc kubenswrapper[4796]: I0930 16:57:41.970874 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6"} err="failed to get container status \"36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6\": rpc error: code = NotFound desc = could not find container \"36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6\": container with ID starting with 36f5a48cc046dc14625ae4c5f5874e8f9b8e3f95af8d99414bfb92c1603a32f6 not found: ID does not exist" Sep 30 16:57:42 crc kubenswrapper[4796]: I0930 16:57:42.744908 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" path="/var/lib/kubelet/pods/58c16fe7-0b61-44f3-b089-4195852e1a09/volumes" Sep 30 16:57:48 crc kubenswrapper[4796]: I0930 16:57:48.916637 4796 generic.go:334] "Generic (PLEG): container finished" podID="e8acd478-67ee-465a-b801-6731a62328f3" containerID="cf53529e111ef1032f81542560dbd2b09f8462543e106d2a5371bdd7fd4bfe1c" exitCode=0 Sep 30 16:57:48 crc kubenswrapper[4796]: I0930 16:57:48.916759 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" event={"ID":"e8acd478-67ee-465a-b801-6731a62328f3","Type":"ContainerDied","Data":"cf53529e111ef1032f81542560dbd2b09f8462543e106d2a5371bdd7fd4bfe1c"} Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.443470 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.541892 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-inventory\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.541975 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-2\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.542159 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ssh-key\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.542229 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-1\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.542308 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-0\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.542338 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pd75\" (UniqueName: \"kubernetes.io/projected/e8acd478-67ee-465a-b801-6731a62328f3-kube-api-access-2pd75\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.542390 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-telemetry-combined-ca-bundle\") pod \"e8acd478-67ee-465a-b801-6731a62328f3\" (UID: \"e8acd478-67ee-465a-b801-6731a62328f3\") " Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.547762 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8acd478-67ee-465a-b801-6731a62328f3-kube-api-access-2pd75" (OuterVolumeSpecName: "kube-api-access-2pd75") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "kube-api-access-2pd75". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.549383 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.572919 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-inventory" (OuterVolumeSpecName: "inventory") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.576116 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.584501 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.588429 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.594677 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "e8acd478-67ee-465a-b801-6731a62328f3" (UID: "e8acd478-67ee-465a-b801-6731a62328f3"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.645491 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.645776 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.645872 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pd75\" (UniqueName: \"kubernetes.io/projected/e8acd478-67ee-465a-b801-6731a62328f3-kube-api-access-2pd75\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.645957 4796 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.646060 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.646166 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.646248 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8acd478-67ee-465a-b801-6731a62328f3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.941646 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" event={"ID":"e8acd478-67ee-465a-b801-6731a62328f3","Type":"ContainerDied","Data":"103dbeadb4570aa925d439ed98844e3042251360153db6a43549e95fdff6feea"} Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.941686 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="103dbeadb4570aa925d439ed98844e3042251360153db6a43549e95fdff6feea" Sep 30 16:57:50 crc kubenswrapper[4796]: I0930 16:57:50.941734 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p" Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.611737 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.613058 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9f48ae38-087f-474e-8e0c-6c737de6da53" containerName="kube-state-metrics" containerID="cri-o://52800a4bc419e5ffb650c14a5a717e854e34a5093aafc814729d1b067b5bbfe5" gracePeriod=30 Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.664794 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.665286 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-central-agent" containerID="cri-o://4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f" gracePeriod=30 Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.666358 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="proxy-httpd" containerID="cri-o://067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3" gracePeriod=30 Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.666473 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="sg-core" containerID="cri-o://6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622" gracePeriod=30 Sep 30 16:57:58 crc kubenswrapper[4796]: I0930 16:57:58.666527 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-notification-agent" containerID="cri-o://d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57" gracePeriod=30 Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.054082 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerID="6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622" exitCode=2 Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.054345 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerDied","Data":"6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622"} Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.104253 4796 generic.go:334] "Generic (PLEG): container finished" podID="9f48ae38-087f-474e-8e0c-6c737de6da53" containerID="52800a4bc419e5ffb650c14a5a717e854e34a5093aafc814729d1b067b5bbfe5" exitCode=2 Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.104302 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9f48ae38-087f-474e-8e0c-6c737de6da53","Type":"ContainerDied","Data":"52800a4bc419e5ffb650c14a5a717e854e34a5093aafc814729d1b067b5bbfe5"} Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.286713 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.423260 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-combined-ca-bundle\") pod \"9f48ae38-087f-474e-8e0c-6c737de6da53\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.423440 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-certs\") pod \"9f48ae38-087f-474e-8e0c-6c737de6da53\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.423871 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svdxf\" (UniqueName: \"kubernetes.io/projected/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-api-access-svdxf\") pod \"9f48ae38-087f-474e-8e0c-6c737de6da53\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.425231 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-config\") pod \"9f48ae38-087f-474e-8e0c-6c737de6da53\" (UID: \"9f48ae38-087f-474e-8e0c-6c737de6da53\") " Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.435191 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-api-access-svdxf" (OuterVolumeSpecName: "kube-api-access-svdxf") pod "9f48ae38-087f-474e-8e0c-6c737de6da53" (UID: "9f48ae38-087f-474e-8e0c-6c737de6da53"). InnerVolumeSpecName "kube-api-access-svdxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.459102 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "9f48ae38-087f-474e-8e0c-6c737de6da53" (UID: "9f48ae38-087f-474e-8e0c-6c737de6da53"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.460699 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f48ae38-087f-474e-8e0c-6c737de6da53" (UID: "9f48ae38-087f-474e-8e0c-6c737de6da53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.485738 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "9f48ae38-087f-474e-8e0c-6c737de6da53" (UID: "9f48ae38-087f-474e-8e0c-6c737de6da53"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.528556 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.528590 4796 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.528602 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svdxf\" (UniqueName: \"kubernetes.io/projected/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-api-access-svdxf\") on node \"crc\" DevicePath \"\"" Sep 30 16:57:59 crc kubenswrapper[4796]: I0930 16:57:59.528611 4796 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9f48ae38-087f-474e-8e0c-6c737de6da53-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.115054 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerID="067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3" exitCode=0 Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.115088 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerID="4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f" exitCode=0 Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.115138 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerDied","Data":"067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3"} Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.115197 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerDied","Data":"4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f"} Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.118094 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9f48ae38-087f-474e-8e0c-6c737de6da53","Type":"ContainerDied","Data":"2ee9b9bca298554a79d75bc277cbe99aac0b4f9855f9c9143d2da788bbad2348"} Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.118133 4796 scope.go:117] "RemoveContainer" containerID="52800a4bc419e5ffb650c14a5a717e854e34a5093aafc814729d1b067b5bbfe5" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.118167 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.152937 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.160626 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.717380 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.755407 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f48ae38-087f-474e-8e0c-6c737de6da53" path="/var/lib/kubelet/pods/9f48ae38-087f-474e-8e0c-6c737de6da53/volumes" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854390 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-log-httpd\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854517 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-scripts\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854570 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-config-data\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854592 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-ceilometer-tls-certs\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854682 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-run-httpd\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854755 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-combined-ca-bundle\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854800 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkgff\" (UniqueName: \"kubernetes.io/projected/cb9a76fb-3656-4a79-96a8-650c99bac0d0-kube-api-access-jkgff\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.854826 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-sg-core-conf-yaml\") pod \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\" (UID: \"cb9a76fb-3656-4a79-96a8-650c99bac0d0\") " Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.855135 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.856641 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.856740 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.861771 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-scripts" (OuterVolumeSpecName: "scripts") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.873261 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb9a76fb-3656-4a79-96a8-650c99bac0d0-kube-api-access-jkgff" (OuterVolumeSpecName: "kube-api-access-jkgff") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "kube-api-access-jkgff". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.909179 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.930915 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.944459 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.959097 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.959131 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.959143 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb9a76fb-3656-4a79-96a8-650c99bac0d0-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.959152 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.959160 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkgff\" (UniqueName: \"kubernetes.io/projected/cb9a76fb-3656-4a79-96a8-650c99bac0d0-kube-api-access-jkgff\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.959168 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:00 crc kubenswrapper[4796]: I0930 16:58:00.995319 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-config-data" (OuterVolumeSpecName: "config-data") pod "cb9a76fb-3656-4a79-96a8-650c99bac0d0" (UID: "cb9a76fb-3656-4a79-96a8-650c99bac0d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.061626 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9a76fb-3656-4a79-96a8-650c99bac0d0-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.096845 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.096910 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.131298 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerID="d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57" exitCode=0 Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.131357 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.131394 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerDied","Data":"d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57"} Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.131447 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb9a76fb-3656-4a79-96a8-650c99bac0d0","Type":"ContainerDied","Data":"9524ad6481c288d48a368551c647202cecaaaa62676fa3f828e4904f514b68ec"} Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.131465 4796 scope.go:117] "RemoveContainer" containerID="067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.163289 4796 scope.go:117] "RemoveContainer" containerID="6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.191865 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.196383 4796 scope.go:117] "RemoveContainer" containerID="d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.203433 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.232916 4796 scope.go:117] "RemoveContainer" containerID="4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.258332 4796 scope.go:117] "RemoveContainer" containerID="067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3" Sep 30 16:58:01 crc kubenswrapper[4796]: E0930 16:58:01.258843 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3\": container with ID starting with 067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3 not found: ID does not exist" containerID="067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.258887 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3"} err="failed to get container status \"067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3\": rpc error: code = NotFound desc = could not find container \"067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3\": container with ID starting with 067c040497293bfc684848c0a88a826346ad0471c5984aba7efbcd2fb329e1b3 not found: ID does not exist" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.258918 4796 scope.go:117] "RemoveContainer" containerID="6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622" Sep 30 16:58:01 crc kubenswrapper[4796]: E0930 16:58:01.259248 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622\": container with ID starting with 6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622 not found: ID does not exist" containerID="6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.259276 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622"} err="failed to get container status \"6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622\": rpc error: code = NotFound desc = could not find container \"6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622\": container with ID starting with 6cbddfeae209da4c5512d3be0bbf277291b5dde89346a9365f67e6740a2d4622 not found: ID does not exist" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.259294 4796 scope.go:117] "RemoveContainer" containerID="d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57" Sep 30 16:58:01 crc kubenswrapper[4796]: E0930 16:58:01.259489 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57\": container with ID starting with d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57 not found: ID does not exist" containerID="d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.259518 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57"} err="failed to get container status \"d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57\": rpc error: code = NotFound desc = could not find container \"d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57\": container with ID starting with d40a17d90e8eab6eccff8436a571969a25b333c9092d5fbd8266b836d2a6bd57 not found: ID does not exist" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.259534 4796 scope.go:117] "RemoveContainer" containerID="4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f" Sep 30 16:58:01 crc kubenswrapper[4796]: E0930 16:58:01.259707 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f\": container with ID starting with 4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f not found: ID does not exist" containerID="4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f" Sep 30 16:58:01 crc kubenswrapper[4796]: I0930 16:58:01.259730 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f"} err="failed to get container status \"4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f\": rpc error: code = NotFound desc = could not find container \"4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f\": container with ID starting with 4e224e748376d118944129b043ae7a99515523c9e99b82b135c25d66fbdad59f not found: ID does not exist" Sep 30 16:58:02 crc kubenswrapper[4796]: I0930 16:58:02.744582 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" path="/var/lib/kubelet/pods/cb9a76fb-3656-4a79-96a8-650c99bac0d0/volumes" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696286 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n"] Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.696808 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="registry-server" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696827 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="registry-server" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.696865 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="proxy-httpd" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696873 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="proxy-httpd" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.696887 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-central-agent" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696896 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-central-agent" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.696912 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="extract-utilities" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696922 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="extract-utilities" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.696933 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="extract-content" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696942 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="extract-content" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.696965 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8acd478-67ee-465a-b801-6731a62328f3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.696975 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8acd478-67ee-465a-b801-6731a62328f3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.697014 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f48ae38-087f-474e-8e0c-6c737de6da53" containerName="kube-state-metrics" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.697022 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f48ae38-087f-474e-8e0c-6c737de6da53" containerName="kube-state-metrics" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.697856 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="sg-core" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.697876 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="sg-core" Sep 30 16:58:03 crc kubenswrapper[4796]: E0930 16:58:03.697890 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-notification-agent" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.697901 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-notification-agent" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698181 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="sg-core" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698202 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-central-agent" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698221 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="ceilometer-notification-agent" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698234 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f48ae38-087f-474e-8e0c-6c737de6da53" containerName="kube-state-metrics" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698244 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c16fe7-0b61-44f3-b089-4195852e1a09" containerName="registry-server" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698256 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9a76fb-3656-4a79-96a8-650c99bac0d0" containerName="proxy-httpd" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.698278 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8acd478-67ee-465a-b801-6731a62328f3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.700009 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.702337 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.741030 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n"] Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.808455 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.808646 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw55b\" (UniqueName: \"kubernetes.io/projected/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-kube-api-access-lw55b\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.808690 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.910831 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw55b\" (UniqueName: \"kubernetes.io/projected/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-kube-api-access-lw55b\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.910909 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.910989 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.912754 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.912778 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:03 crc kubenswrapper[4796]: I0930 16:58:03.933781 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw55b\" (UniqueName: \"kubernetes.io/projected/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-kube-api-access-lw55b\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:04 crc kubenswrapper[4796]: I0930 16:58:04.042218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:04 crc kubenswrapper[4796]: I0930 16:58:04.519312 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n"] Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.184252 4796 generic.go:334] "Generic (PLEG): container finished" podID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerID="59cb7c197034143ccb7dca685b34f848faa2007613e160463f4e152f5338f750" exitCode=0 Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.184330 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" event={"ID":"5368fc26-bf5c-4c4d-add0-46d7f1fc0346","Type":"ContainerDied","Data":"59cb7c197034143ccb7dca685b34f848faa2007613e160463f4e152f5338f750"} Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.184582 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" event={"ID":"5368fc26-bf5c-4c4d-add0-46d7f1fc0346","Type":"ContainerStarted","Data":"66525bb84e212b134c23e7863b5778bbc28b60d3296bbfd07cbfbc63a7423012"} Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.410623 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.410869 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="411c488d-d370-41c8-93e1-13c6721eb41d" containerName="openstackclient" containerID="cri-o://f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86" gracePeriod=2 Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.423375 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.453324 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 30 16:58:05 crc kubenswrapper[4796]: E0930 16:58:05.453743 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="411c488d-d370-41c8-93e1-13c6721eb41d" containerName="openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.453769 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="411c488d-d370-41c8-93e1-13c6721eb41d" containerName="openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.455891 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="411c488d-d370-41c8-93e1-13c6721eb41d" containerName="openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.456566 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.462035 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="411c488d-d370-41c8-93e1-13c6721eb41d" podUID="1df486ab-b2b8-471d-b88c-c800f254716c" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.467147 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.548219 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1df486ab-b2b8-471d-b88c-c800f254716c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.548290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1df486ab-b2b8-471d-b88c-c800f254716c-openstack-config-secret\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.548556 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx6pv\" (UniqueName: \"kubernetes.io/projected/1df486ab-b2b8-471d-b88c-c800f254716c-kube-api-access-qx6pv\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.548704 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1df486ab-b2b8-471d-b88c-c800f254716c-openstack-config\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.652680 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1df486ab-b2b8-471d-b88c-c800f254716c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.652791 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1df486ab-b2b8-471d-b88c-c800f254716c-openstack-config-secret\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.652883 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx6pv\" (UniqueName: \"kubernetes.io/projected/1df486ab-b2b8-471d-b88c-c800f254716c-kube-api-access-qx6pv\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.652957 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1df486ab-b2b8-471d-b88c-c800f254716c-openstack-config\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.654001 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1df486ab-b2b8-471d-b88c-c800f254716c-openstack-config\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.680729 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1df486ab-b2b8-471d-b88c-c800f254716c-openstack-config-secret\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.681352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1df486ab-b2b8-471d-b88c-c800f254716c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.684476 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx6pv\" (UniqueName: \"kubernetes.io/projected/1df486ab-b2b8-471d-b88c-c800f254716c-kube-api-access-qx6pv\") pod \"openstackclient\" (UID: \"1df486ab-b2b8-471d-b88c-c800f254716c\") " pod="openstack/openstackclient" Sep 30 16:58:05 crc kubenswrapper[4796]: I0930 16:58:05.776856 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.156692 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-rpkc5"] Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.158444 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.175423 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-rpkc5"] Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.268710 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ll4q\" (UniqueName: \"kubernetes.io/projected/7eb4171f-878b-4a0e-8852-99700fedd6ea-kube-api-access-6ll4q\") pod \"cloudkitty-db-create-rpkc5\" (UID: \"7eb4171f-878b-4a0e-8852-99700fedd6ea\") " pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.361361 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.370790 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ll4q\" (UniqueName: \"kubernetes.io/projected/7eb4171f-878b-4a0e-8852-99700fedd6ea-kube-api-access-6ll4q\") pod \"cloudkitty-db-create-rpkc5\" (UID: \"7eb4171f-878b-4a0e-8852-99700fedd6ea\") " pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.389726 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ll4q\" (UniqueName: \"kubernetes.io/projected/7eb4171f-878b-4a0e-8852-99700fedd6ea-kube-api-access-6ll4q\") pod \"cloudkitty-db-create-rpkc5\" (UID: \"7eb4171f-878b-4a0e-8852-99700fedd6ea\") " pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.482020 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:06 crc kubenswrapper[4796]: I0930 16:58:06.899340 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-rpkc5"] Sep 30 16:58:06 crc kubenswrapper[4796]: W0930 16:58:06.919055 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb4171f_878b_4a0e_8852_99700fedd6ea.slice/crio-f2aad660b718843c05cce6a7424ae7903081c62112e9199fd26b661ad1c96028 WatchSource:0}: Error finding container f2aad660b718843c05cce6a7424ae7903081c62112e9199fd26b661ad1c96028: Status 404 returned error can't find the container with id f2aad660b718843c05cce6a7424ae7903081c62112e9199fd26b661ad1c96028 Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.204245 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1df486ab-b2b8-471d-b88c-c800f254716c","Type":"ContainerStarted","Data":"24d8a36f4209df701c0450e19652e56fb73d209623788497e22da604935ecf6a"} Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.204566 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1df486ab-b2b8-471d-b88c-c800f254716c","Type":"ContainerStarted","Data":"27be4011cb1c5752518c1119987361380e6679006e593ac18f193eba17c1209d"} Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.205931 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-rpkc5" event={"ID":"7eb4171f-878b-4a0e-8852-99700fedd6ea","Type":"ContainerStarted","Data":"200c049deea2c36491d4c247c332f4cdf2e66a0821aff24af4c41c16cdb95f14"} Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.205999 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-rpkc5" event={"ID":"7eb4171f-878b-4a0e-8852-99700fedd6ea","Type":"ContainerStarted","Data":"f2aad660b718843c05cce6a7424ae7903081c62112e9199fd26b661ad1c96028"} Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.207675 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" event={"ID":"5368fc26-bf5c-4c4d-add0-46d7f1fc0346","Type":"ContainerStarted","Data":"d038c51fc0fd93218b9c0190a192c3f4be9fdb87aaadac1e2bc79c85abad0163"} Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.221510 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.221492329 podStartE2EDuration="2.221492329s" podCreationTimestamp="2025-09-30 16:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:58:07.218575767 +0000 UTC m=+2779.231854294" watchObservedRunningTime="2025-09-30 16:58:07.221492329 +0000 UTC m=+2779.234770856" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.264919 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-create-rpkc5" podStartSLOduration=1.2648989290000001 podStartE2EDuration="1.264898929s" podCreationTimestamp="2025-09-30 16:58:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:58:07.25741881 +0000 UTC m=+2779.270697337" watchObservedRunningTime="2025-09-30 16:58:07.264898929 +0000 UTC m=+2779.278177616" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.700528 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.702572 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.704771 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.714622 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-28x5n" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.714909 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.715097 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.735401 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801125 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjmwg\" (UniqueName: \"kubernetes.io/projected/411c488d-d370-41c8-93e1-13c6721eb41d-kube-api-access-mjmwg\") pod \"411c488d-d370-41c8-93e1-13c6721eb41d\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801192 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config\") pod \"411c488d-d370-41c8-93e1-13c6721eb41d\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801265 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config-secret\") pod \"411c488d-d370-41c8-93e1-13c6721eb41d\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801429 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-combined-ca-bundle\") pod \"411c488d-d370-41c8-93e1-13c6721eb41d\" (UID: \"411c488d-d370-41c8-93e1-13c6721eb41d\") " Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801690 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-log-httpd\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801748 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqpz6\" (UniqueName: \"kubernetes.io/projected/e536dcbd-a48d-409a-8154-49e9e392f3b4-kube-api-access-qqpz6\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801805 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-config-data\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801846 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-scripts\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.801861 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-run-httpd\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.823057 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/411c488d-d370-41c8-93e1-13c6721eb41d-kube-api-access-mjmwg" (OuterVolumeSpecName: "kube-api-access-mjmwg") pod "411c488d-d370-41c8-93e1-13c6721eb41d" (UID: "411c488d-d370-41c8-93e1-13c6721eb41d"). InnerVolumeSpecName "kube-api-access-mjmwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.836872 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "411c488d-d370-41c8-93e1-13c6721eb41d" (UID: "411c488d-d370-41c8-93e1-13c6721eb41d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.878938 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "411c488d-d370-41c8-93e1-13c6721eb41d" (UID: "411c488d-d370-41c8-93e1-13c6721eb41d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.886379 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "411c488d-d370-41c8-93e1-13c6721eb41d" (UID: "411c488d-d370-41c8-93e1-13c6721eb41d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903392 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-log-httpd\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903463 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqpz6\" (UniqueName: \"kubernetes.io/projected/e536dcbd-a48d-409a-8154-49e9e392f3b4-kube-api-access-qqpz6\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903537 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-config-data\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903580 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-scripts\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903594 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-run-httpd\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903672 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903683 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjmwg\" (UniqueName: \"kubernetes.io/projected/411c488d-d370-41c8-93e1-13c6721eb41d-kube-api-access-mjmwg\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903694 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.903702 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/411c488d-d370-41c8-93e1-13c6721eb41d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.904101 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-log-httpd\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.905050 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-run-httpd\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.908631 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-config-data\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.908738 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-scripts\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.909024 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:07 crc kubenswrapper[4796]: I0930 16:58:07.929371 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqpz6\" (UniqueName: \"kubernetes.io/projected/e536dcbd-a48d-409a-8154-49e9e392f3b4-kube-api-access-qqpz6\") pod \"ceilometer-0\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " pod="openstack/ceilometer-0" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.059276 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.222389 4796 generic.go:334] "Generic (PLEG): container finished" podID="411c488d-d370-41c8-93e1-13c6721eb41d" containerID="f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86" exitCode=137 Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.222683 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.222717 4796 scope.go:117] "RemoveContainer" containerID="f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.227673 4796 generic.go:334] "Generic (PLEG): container finished" podID="7eb4171f-878b-4a0e-8852-99700fedd6ea" containerID="200c049deea2c36491d4c247c332f4cdf2e66a0821aff24af4c41c16cdb95f14" exitCode=0 Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.227758 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-rpkc5" event={"ID":"7eb4171f-878b-4a0e-8852-99700fedd6ea","Type":"ContainerDied","Data":"200c049deea2c36491d4c247c332f4cdf2e66a0821aff24af4c41c16cdb95f14"} Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.233041 4796 generic.go:334] "Generic (PLEG): container finished" podID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerID="d038c51fc0fd93218b9c0190a192c3f4be9fdb87aaadac1e2bc79c85abad0163" exitCode=0 Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.233129 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" event={"ID":"5368fc26-bf5c-4c4d-add0-46d7f1fc0346","Type":"ContainerDied","Data":"d038c51fc0fd93218b9c0190a192c3f4be9fdb87aaadac1e2bc79c85abad0163"} Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.252160 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="411c488d-d370-41c8-93e1-13c6721eb41d" podUID="1df486ab-b2b8-471d-b88c-c800f254716c" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.267792 4796 scope.go:117] "RemoveContainer" containerID="f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86" Sep 30 16:58:08 crc kubenswrapper[4796]: E0930 16:58:08.274297 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86\": container with ID starting with f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86 not found: ID does not exist" containerID="f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.274349 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86"} err="failed to get container status \"f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86\": rpc error: code = NotFound desc = could not find container \"f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86\": container with ID starting with f046be362f82c7aa1a6bfb03f0c95546584751922ddbf4f0266288b4cf039d86 not found: ID does not exist" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.317863 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq"] Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.319871 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.336295 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq"] Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.514309 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xxtd\" (UniqueName: \"kubernetes.io/projected/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-kube-api-access-2xxtd\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.514687 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.514759 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.588450 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 16:58:08 crc kubenswrapper[4796]: W0930 16:58:08.605910 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode536dcbd_a48d_409a_8154_49e9e392f3b4.slice/crio-11d8abd3345cf2814d4027c78bffa5e62cdc0e21792b584e26de4039079c00dc WatchSource:0}: Error finding container 11d8abd3345cf2814d4027c78bffa5e62cdc0e21792b584e26de4039079c00dc: Status 404 returned error can't find the container with id 11d8abd3345cf2814d4027c78bffa5e62cdc0e21792b584e26de4039079c00dc Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.616696 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xxtd\" (UniqueName: \"kubernetes.io/projected/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-kube-api-access-2xxtd\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.616762 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.616819 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.617352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.617821 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.644284 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xxtd\" (UniqueName: \"kubernetes.io/projected/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-kube-api-access-2xxtd\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.656133 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:08 crc kubenswrapper[4796]: I0930 16:58:08.908080 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="411c488d-d370-41c8-93e1-13c6721eb41d" path="/var/lib/kubelet/pods/411c488d-d370-41c8-93e1-13c6721eb41d/volumes" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.242083 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerStarted","Data":"11d8abd3345cf2814d4027c78bffa5e62cdc0e21792b584e26de4039079c00dc"} Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.244100 4796 generic.go:334] "Generic (PLEG): container finished" podID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerID="23b2e7c2a234c90c3bef55917bcc303ecb1246781c87ad97484345e719cb653f" exitCode=0 Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.244145 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" event={"ID":"5368fc26-bf5c-4c4d-add0-46d7f1fc0346","Type":"ContainerDied","Data":"23b2e7c2a234c90c3bef55917bcc303ecb1246781c87ad97484345e719cb653f"} Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.309679 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq"] Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.852638 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9"] Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.854288 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.858559 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.858650 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-kp6hb" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.858840 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.858939 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.859167 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.869334 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9"] Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.949099 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.949398 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e35857-aed6-49cf-8ede-715b5b915ffe-config\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.949493 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vzgc\" (UniqueName: \"kubernetes.io/projected/e8e35857-aed6-49cf-8ede-715b5b915ffe-kube-api-access-8vzgc\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.949607 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:09 crc kubenswrapper[4796]: I0930 16:58:09.949908 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.051727 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vzgc\" (UniqueName: \"kubernetes.io/projected/e8e35857-aed6-49cf-8ede-715b5b915ffe-kube-api-access-8vzgc\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.051791 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.051885 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.051912 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.052045 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e35857-aed6-49cf-8ede-715b5b915ffe-config\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.052960 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e35857-aed6-49cf-8ede-715b5b915ffe-config\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.055369 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.058761 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.078661 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/e8e35857-aed6-49cf-8ede-715b5b915ffe-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.083301 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vzgc\" (UniqueName: \"kubernetes.io/projected/e8e35857-aed6-49cf-8ede-715b5b915ffe-kube-api-access-8vzgc\") pod \"cloudkitty-lokistack-distributor-bccccd5f6-j4nk9\" (UID: \"e8e35857-aed6-49cf-8ede-715b5b915ffe\") " pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.125307 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.126684 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.128569 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-loki-s3" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.128829 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.130952 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.144206 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.176534 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.256155 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.256238 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.256283 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bv82\" (UniqueName: \"kubernetes.io/projected/914424ae-a0ec-4578-9a57-824fceadcf0b-kube-api-access-5bv82\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.256304 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-logging-loki-s3\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.256330 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/914424ae-a0ec-4578-9a57-824fceadcf0b-config\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.256347 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.260014 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.261336 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.265330 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.265583 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.300046 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.304406 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.322879 4796 generic.go:334] "Generic (PLEG): container finished" podID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerID="21ccaebe713abdbf42b1c5c144c314458bc6dd9adf536be2136fac3158859f56" exitCode=0 Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.322885 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" event={"ID":"0d843fc8-0ae6-4498-97b4-9b40bafd59d0","Type":"ContainerDied","Data":"21ccaebe713abdbf42b1c5c144c314458bc6dd9adf536be2136fac3158859f56"} Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.322956 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" event={"ID":"0d843fc8-0ae6-4498-97b4-9b40bafd59d0","Type":"ContainerStarted","Data":"a4dc64a54d74f4508f7338fe8ec7ab8e024a31487643c525427141d13c6b288d"} Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.333432 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-rpkc5" event={"ID":"7eb4171f-878b-4a0e-8852-99700fedd6ea","Type":"ContainerDied","Data":"f2aad660b718843c05cce6a7424ae7903081c62112e9199fd26b661ad1c96028"} Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.333470 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2aad660b718843c05cce6a7424ae7903081c62112e9199fd26b661ad1c96028" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.333521 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-rpkc5" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.342458 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerStarted","Data":"59877affe9f0dcd0279809ecef684a1b9996fd17cc7da6da6dd7ec8ec67e2e88"} Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.357608 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ll4q\" (UniqueName: \"kubernetes.io/projected/7eb4171f-878b-4a0e-8852-99700fedd6ea-kube-api-access-6ll4q\") pod \"7eb4171f-878b-4a0e-8852-99700fedd6ea\" (UID: \"7eb4171f-878b-4a0e-8852-99700fedd6ea\") " Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.358647 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.358795 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.358862 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.359003 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360096 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360232 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bv82\" (UniqueName: \"kubernetes.io/projected/914424ae-a0ec-4578-9a57-824fceadcf0b-kube-api-access-5bv82\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360273 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-logging-loki-s3\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360322 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18114f1-3f03-4f73-b46e-62056032d351-config\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360355 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/914424ae-a0ec-4578-9a57-824fceadcf0b-config\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360384 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.360438 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwftc\" (UniqueName: \"kubernetes.io/projected/b18114f1-3f03-4f73-b46e-62056032d351-kube-api-access-gwftc\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.361064 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.361719 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/914424ae-a0ec-4578-9a57-824fceadcf0b-config\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.369545 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb4171f-878b-4a0e-8852-99700fedd6ea-kube-api-access-6ll4q" (OuterVolumeSpecName: "kube-api-access-6ll4q") pod "7eb4171f-878b-4a0e-8852-99700fedd6ea" (UID: "7eb4171f-878b-4a0e-8852-99700fedd6ea"). InnerVolumeSpecName "kube-api-access-6ll4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.372704 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.373620 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.374006 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/914424ae-a0ec-4578-9a57-824fceadcf0b-logging-loki-s3\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.407940 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bv82\" (UniqueName: \"kubernetes.io/projected/914424ae-a0ec-4578-9a57-824fceadcf0b-kube-api-access-5bv82\") pod \"cloudkitty-lokistack-querier-6b6cdc96db-2rjz6\" (UID: \"914424ae-a0ec-4578-9a57-824fceadcf0b\") " pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.413806 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x"] Sep 30 16:58:10 crc kubenswrapper[4796]: E0930 16:58:10.414257 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb4171f-878b-4a0e-8852-99700fedd6ea" containerName="mariadb-database-create" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.414274 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb4171f-878b-4a0e-8852-99700fedd6ea" containerName="mariadb-database-create" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.414463 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb4171f-878b-4a0e-8852-99700fedd6ea" containerName="mariadb-database-create" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.415218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.425580 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.425629 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.427411 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.427601 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.427969 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-n7m56" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.428266 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.428512 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.461952 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.463644 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18114f1-3f03-4f73-b46e-62056032d351-config\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464693 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwq65\" (UniqueName: \"kubernetes.io/projected/165eb10b-4c60-4635-928e-c04a72e9b4f5-kube-api-access-bwq65\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464722 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwftc\" (UniqueName: \"kubernetes.io/projected/b18114f1-3f03-4f73-b46e-62056032d351-kube-api-access-gwftc\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464755 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464795 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464844 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464871 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-rbac\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464894 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464916 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464951 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.464992 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.465044 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.465080 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tenants\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.465105 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.465148 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ll4q\" (UniqueName: \"kubernetes.io/projected/7eb4171f-878b-4a0e-8852-99700fedd6ea-kube-api-access-6ll4q\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.465991 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18114f1-3f03-4f73-b46e-62056032d351-config\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.468433 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.483875 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwftc\" (UniqueName: \"kubernetes.io/projected/b18114f1-3f03-4f73-b46e-62056032d351-kube-api-access-gwftc\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.484273 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.490048 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.504613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b18114f1-3f03-4f73-b46e-62056032d351-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f\" (UID: \"b18114f1-3f03-4f73-b46e-62056032d351\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.512128 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn"] Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.527790 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578453 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578512 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-rbac\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578537 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578583 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578605 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-rbac\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578633 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578662 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578793 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578823 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tenants\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578843 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578885 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578915 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwq65\" (UniqueName: \"kubernetes.io/projected/165eb10b-4c60-4635-928e-c04a72e9b4f5-kube-api-access-bwq65\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.578985 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.579010 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.579028 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-tenants\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.579057 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.579072 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhw69\" (UniqueName: \"kubernetes.io/projected/15e69566-5b25-4929-a40b-1a3e2d710800-kube-api-access-jhw69\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.579503 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.580196 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.580784 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-rbac\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: E0930 16:58:10.580859 4796 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Sep 30 16:58:10 crc kubenswrapper[4796]: E0930 16:58:10.580901 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tls-secret podName:165eb10b-4c60-4635-928e-c04a72e9b4f5 nodeName:}" failed. No retries permitted until 2025-09-30 16:58:11.080886082 +0000 UTC m=+2783.094164599 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tls-secret") pod "cloudkitty-lokistack-gateway-89dc74b89-cpv5x" (UID: "165eb10b-4c60-4635-928e-c04a72e9b4f5") : secret "cloudkitty-lokistack-gateway-http" not found Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.581500 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.582521 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.582670 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/165eb10b-4c60-4635-928e-c04a72e9b4f5-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.589494 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tenants\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.597841 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwq65\" (UniqueName: \"kubernetes.io/projected/165eb10b-4c60-4635-928e-c04a72e9b4f5-kube-api-access-bwq65\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.598172 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.683695 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.684194 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.685219 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691328 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691494 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-tenants\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691660 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691688 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhw69\" (UniqueName: \"kubernetes.io/projected/15e69566-5b25-4929-a40b-1a3e2d710800-kube-api-access-jhw69\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.691861 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-rbac\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.693191 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-rbac\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.696337 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.696893 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.701743 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.710493 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15e69566-5b25-4929-a40b-1a3e2d710800-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.712701 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-tenants\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.717580 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/15e69566-5b25-4929-a40b-1a3e2d710800-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.724660 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhw69\" (UniqueName: \"kubernetes.io/projected/15e69566-5b25-4929-a40b-1a3e2d710800-kube-api-access-jhw69\") pod \"cloudkitty-lokistack-gateway-89dc74b89-hv6jn\" (UID: \"15e69566-5b25-4929-a40b-1a3e2d710800\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:10 crc kubenswrapper[4796]: I0930 16:58:10.795746 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.032621 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.100794 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-util\") pod \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.100859 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.100878 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-bundle\") pod \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.100919 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw55b\" (UniqueName: \"kubernetes.io/projected/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-kube-api-access-lw55b\") pod \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\" (UID: \"5368fc26-bf5c-4c4d-add0-46d7f1fc0346\") " Sep 30 16:58:11 crc kubenswrapper[4796]: E0930 16:58:11.101295 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="extract" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.101308 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="extract" Sep 30 16:58:11 crc kubenswrapper[4796]: E0930 16:58:11.101325 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="pull" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.101331 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="pull" Sep 30 16:58:11 crc kubenswrapper[4796]: E0930 16:58:11.101353 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="util" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.101360 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="util" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.101360 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.101541 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5368fc26-bf5c-4c4d-add0-46d7f1fc0346" containerName="extract" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.102218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.104320 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-bundle" (OuterVolumeSpecName: "bundle") pod "5368fc26-bf5c-4c4d-add0-46d7f1fc0346" (UID: "5368fc26-bf5c-4c4d-add0-46d7f1fc0346"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.108699 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.108907 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.109549 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/165eb10b-4c60-4635-928e-c04a72e9b4f5-tls-secret\") pod \"cloudkitty-lokistack-gateway-89dc74b89-cpv5x\" (UID: \"165eb10b-4c60-4635-928e-c04a72e9b4f5\") " pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.111458 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.112139 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-kube-api-access-lw55b" (OuterVolumeSpecName: "kube-api-access-lw55b") pod "5368fc26-bf5c-4c4d-add0-46d7f1fc0346" (UID: "5368fc26-bf5c-4c4d-add0-46d7f1fc0346"). InnerVolumeSpecName "kube-api-access-lw55b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.122299 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-util" (OuterVolumeSpecName: "util") pod "5368fc26-bf5c-4c4d-add0-46d7f1fc0346" (UID: "5368fc26-bf5c-4c4d-add0-46d7f1fc0346"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.202769 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.202838 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.202876 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.202910 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-logging-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.202961 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd3fc7-5956-4320-b074-bc791068c9cc-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.203005 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkqln\" (UniqueName: \"kubernetes.io/projected/6cbd3fc7-5956-4320-b074-bc791068c9cc-kube-api-access-vkqln\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.203153 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.203226 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.203321 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.203341 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.203355 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw55b\" (UniqueName: \"kubernetes.io/projected/5368fc26-bf5c-4c4d-add0-46d7f1fc0346-kube-api-access-lw55b\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.207229 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.230439 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9"] Sep 30 16:58:11 crc kubenswrapper[4796]: W0930 16:58:11.241103 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8e35857_aed6_49cf_8ede_715b5b915ffe.slice/crio-b5652c920caedf6296307a1136a189c028ec3ad4edb934c846a02dad86e863ed WatchSource:0}: Error finding container b5652c920caedf6296307a1136a189c028ec3ad4edb934c846a02dad86e863ed: Status 404 returned error can't find the container with id b5652c920caedf6296307a1136a189c028ec3ad4edb934c846a02dad86e863ed Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304781 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304826 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304848 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304900 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-logging-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304937 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd3fc7-5956-4320-b074-bc791068c9cc-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.304956 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkqln\" (UniqueName: \"kubernetes.io/projected/6cbd3fc7-5956-4320-b074-bc791068c9cc-kube-api-access-vkqln\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.305882 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.306271 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd3fc7-5956-4320-b074-bc791068c9cc-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.311599 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.314077 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.317370 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6cbd3fc7-5956-4320-b074-bc791068c9cc-logging-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.320923 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.320992 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/23369794d8239f6f859421d371662fc8059eb0c8edbc527866d7ae6ac47bcce0/globalmount\"" pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.321015 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.321048 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e522a7bc2b0d52c95e9cb1d5458c364bd6904a6afa156b28e3659eff332a4ff0/globalmount\"" pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.330520 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkqln\" (UniqueName: \"kubernetes.io/projected/6cbd3fc7-5956-4320-b074-bc791068c9cc-kube-api-access-vkqln\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.358398 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerStarted","Data":"805681bc59c000c3d10eafcf08305f67ed12c5d256a6d8ffbb4eb0a1a6d7a2ab"} Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.358460 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.360366 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.373447 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.373623 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.374543 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.385768 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.388387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" event={"ID":"5368fc26-bf5c-4c4d-add0-46d7f1fc0346","Type":"ContainerDied","Data":"66525bb84e212b134c23e7863b5778bbc28b60d3296bbfd07cbfbc63a7423012"} Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.388416 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66525bb84e212b134c23e7863b5778bbc28b60d3296bbfd07cbfbc63a7423012" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.388480 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.400502 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" event={"ID":"914424ae-a0ec-4578-9a57-824fceadcf0b","Type":"ContainerStarted","Data":"e0cd003a6fec67ea135e5694ce390a32699661ad6a3d92045221271fb1758472"} Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.407652 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-logging-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.407708 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.407782 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.407812 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.408032 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.408078 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.408142 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsngq\" (UniqueName: \"kubernetes.io/projected/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-kube-api-access-gsngq\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.411471 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.413238 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" event={"ID":"e8e35857-aed6-49cf-8ede-715b5b915ffe","Type":"ContainerStarted","Data":"b5652c920caedf6296307a1136a189c028ec3ad4edb934c846a02dad86e863ed"} Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.418653 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aec30a7e-cff4-4cf8-8fdd-1f6e6f32fdda\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.418897 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f3ebe19-4d4e-4482-8fc6-154a2019ecf4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"6cbd3fc7-5956-4320-b074-bc791068c9cc\") " pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.437294 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.498747 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn"] Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.510714 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.510897 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.510950 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsngq\" (UniqueName: \"kubernetes.io/projected/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-kube-api-access-gsngq\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.511017 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-logging-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.511041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.511079 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.511102 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.514842 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.516575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.517618 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.522930 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.522962 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/86a5cc0efe1e561a9f99cf5f449cbe28420090449875dfbcd8b11535d8e302c3/globalmount\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.523813 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-logging-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.524234 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.548360 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsngq\" (UniqueName: \"kubernetes.io/projected/048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c-kube-api-access-gsngq\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.580828 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b71e8d6-e561-40e2-b424-48456e4eb125\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:11 crc kubenswrapper[4796]: I0930 16:58:11.731586 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.051167 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x"] Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.288575 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Sep 30 16:58:12 crc kubenswrapper[4796]: W0930 16:58:12.373105 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod165eb10b_4c60_4635_928e_c04a72e9b4f5.slice/crio-26c409389898e5834c6704ffb2913828cf5521738addf0957764ee5b1b42ccd0 WatchSource:0}: Error finding container 26c409389898e5834c6704ffb2913828cf5521738addf0957764ee5b1b42ccd0: Status 404 returned error can't find the container with id 26c409389898e5834c6704ffb2913828cf5521738addf0957764ee5b1b42ccd0 Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.443679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"6cbd3fc7-5956-4320-b074-bc791068c9cc","Type":"ContainerStarted","Data":"17abc7217723a6fb8654d090e5c9f32171b188980b7751d72c0e487af1b278d4"} Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.445774 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.456413 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerStarted","Data":"a5b6f93acd32c79f8c4eb28dc6a1b8676f1ff717f6fcebb8a524e70d9efc209a"} Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.464409 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerStarted","Data":"42a6b5cb9d6b4b85fe799eedf94866ca97de90de1b559246f517c4346e7e041d"} Sep 30 16:58:12 crc kubenswrapper[4796]: W0930 16:58:12.466485 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod048ee9b2_bfc7_4f0c_ae5b_1fab51d7dc0c.slice/crio-341bf9f08e2b3ab6827c2bf29c5a63bc8d5ef41f5f83fcfe21935702d52ee051 WatchSource:0}: Error finding container 341bf9f08e2b3ab6827c2bf29c5a63bc8d5ef41f5f83fcfe21935702d52ee051: Status 404 returned error can't find the container with id 341bf9f08e2b3ab6827c2bf29c5a63bc8d5ef41f5f83fcfe21935702d52ee051 Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.467742 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerStarted","Data":"26c409389898e5834c6704ffb2913828cf5521738addf0957764ee5b1b42ccd0"} Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.471250 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" event={"ID":"b18114f1-3f03-4f73-b46e-62056032d351","Type":"ContainerStarted","Data":"5bf620d7831aa5da57fc23145116c388612a10e7a4fdaba3ce47c50c3167f457"} Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.504859 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.506342 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.509510 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.509548 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.516994 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646201 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e12debc-3efb-4e67-9e07-2531628e41ab-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s755j\" (UniqueName: \"kubernetes.io/projected/0e12debc-3efb-4e67-9e07-2531628e41ab-kube-api-access-s755j\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646292 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646339 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646361 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-logging-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646386 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.646415 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.748274 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.748736 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e12debc-3efb-4e67-9e07-2531628e41ab-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.748784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s755j\" (UniqueName: \"kubernetes.io/projected/0e12debc-3efb-4e67-9e07-2531628e41ab-kube-api-access-s755j\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.748855 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.749025 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.749769 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.749836 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-logging-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.749845 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e12debc-3efb-4e67-9e07-2531628e41ab-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.749867 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.755224 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-logging-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.755304 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.755338 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/31f8ee63022ca6bf003134a2787fe0ce00bc401c1052674f10989c4fa1cf8a6e/globalmount\"" pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.755641 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.757401 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/0e12debc-3efb-4e67-9e07-2531628e41ab-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.772568 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s755j\" (UniqueName: \"kubernetes.io/projected/0e12debc-3efb-4e67-9e07-2531628e41ab-kube-api-access-s755j\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.799088 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef440b99-e396-43bd-8e90-d8c14f1e18fc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"0e12debc-3efb-4e67-9e07-2531628e41ab\") " pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:12 crc kubenswrapper[4796]: I0930 16:58:12.829494 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:13 crc kubenswrapper[4796]: I0930 16:58:13.358580 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Sep 30 16:58:13 crc kubenswrapper[4796]: I0930 16:58:13.496676 4796 generic.go:334] "Generic (PLEG): container finished" podID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerID="42563b517a43dcbee43b6069c7762b4280aabb16d82db0c7dde67d02802766e7" exitCode=0 Sep 30 16:58:13 crc kubenswrapper[4796]: I0930 16:58:13.497353 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" event={"ID":"0d843fc8-0ae6-4498-97b4-9b40bafd59d0","Type":"ContainerDied","Data":"42563b517a43dcbee43b6069c7762b4280aabb16d82db0c7dde67d02802766e7"} Sep 30 16:58:13 crc kubenswrapper[4796]: I0930 16:58:13.504218 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"0e12debc-3efb-4e67-9e07-2531628e41ab","Type":"ContainerStarted","Data":"327783a05044400076fab51b6eab9b9d54dbe2e0848ea28426861b5017c8c32d"} Sep 30 16:58:13 crc kubenswrapper[4796]: I0930 16:58:13.508664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c","Type":"ContainerStarted","Data":"341bf9f08e2b3ab6827c2bf29c5a63bc8d5ef41f5f83fcfe21935702d52ee051"} Sep 30 16:58:14 crc kubenswrapper[4796]: I0930 16:58:14.522958 4796 generic.go:334] "Generic (PLEG): container finished" podID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerID="2f81879468490186d31638af25ae6ad93849a185de4c38647dd6beb679bcf378" exitCode=0 Sep 30 16:58:14 crc kubenswrapper[4796]: I0930 16:58:14.522997 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" event={"ID":"0d843fc8-0ae6-4498-97b4-9b40bafd59d0","Type":"ContainerDied","Data":"2f81879468490186d31638af25ae6ad93849a185de4c38647dd6beb679bcf378"} Sep 30 16:58:17 crc kubenswrapper[4796]: I0930 16:58:17.953032 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.108383 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-util\") pod \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.108562 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xxtd\" (UniqueName: \"kubernetes.io/projected/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-kube-api-access-2xxtd\") pod \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.108632 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-bundle\") pod \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\" (UID: \"0d843fc8-0ae6-4498-97b4-9b40bafd59d0\") " Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.110735 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-bundle" (OuterVolumeSpecName: "bundle") pod "0d843fc8-0ae6-4498-97b4-9b40bafd59d0" (UID: "0d843fc8-0ae6-4498-97b4-9b40bafd59d0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.115189 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-kube-api-access-2xxtd" (OuterVolumeSpecName: "kube-api-access-2xxtd") pod "0d843fc8-0ae6-4498-97b4-9b40bafd59d0" (UID: "0d843fc8-0ae6-4498-97b4-9b40bafd59d0"). InnerVolumeSpecName "kube-api-access-2xxtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.121735 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-util" (OuterVolumeSpecName: "util") pod "0d843fc8-0ae6-4498-97b4-9b40bafd59d0" (UID: "0d843fc8-0ae6-4498-97b4-9b40bafd59d0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.211834 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-util\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.211874 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xxtd\" (UniqueName: \"kubernetes.io/projected/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-kube-api-access-2xxtd\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.211886 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d843fc8-0ae6-4498-97b4-9b40bafd59d0-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.593100 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" event={"ID":"0d843fc8-0ae6-4498-97b4-9b40bafd59d0","Type":"ContainerDied","Data":"a4dc64a54d74f4508f7338fe8ec7ab8e024a31487643c525427141d13c6b288d"} Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.593375 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4dc64a54d74f4508f7338fe8ec7ab8e024a31487643c525427141d13c6b288d" Sep 30 16:58:18 crc kubenswrapper[4796]: I0930 16:58:18.593420 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.626482 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerStarted","Data":"043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.627232 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.628684 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="031e851cc11a95f4d26921d631b298c75748344570bbc5d1e8915f6ae36b9906" exitCode=2 Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.628751 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"031e851cc11a95f4d26921d631b298c75748344570bbc5d1e8915f6ae36b9906"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.629792 4796 scope.go:117] "RemoveContainer" containerID="031e851cc11a95f4d26921d631b298c75748344570bbc5d1e8915f6ae36b9906" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.630273 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"0e12debc-3efb-4e67-9e07-2531628e41ab","Type":"ContainerStarted","Data":"da62c1a20bc1f65dc5c395736cf92bbafddc75ffec3cb2bdeee4120c0166fffc"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.630402 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.632183 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c","Type":"ContainerStarted","Data":"12ed18305a2d599f9653a1d00e4a1ccf62cbab226588cf7f1b7fe3dbed30ba31"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.632509 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.643913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"6cbd3fc7-5956-4320-b074-bc791068c9cc","Type":"ContainerStarted","Data":"0ff7ea03ccf41f0c2f77fdf2177ba15c8ac55e95cecd12d409d2e45a7afe20c9"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.644156 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.646076 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="40c72b999a335d12493390eabdd70eb3010b0ce7fc570702bde45f93b44fb01d" exitCode=2 Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.646138 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"40c72b999a335d12493390eabdd70eb3010b0ce7fc570702bde45f93b44fb01d"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.646527 4796 scope.go:117] "RemoveContainer" containerID="40c72b999a335d12493390eabdd70eb3010b0ce7fc570702bde45f93b44fb01d" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.649225 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" event={"ID":"e8e35857-aed6-49cf-8ede-715b5b915ffe","Type":"ContainerStarted","Data":"5a465b7c35b5ab6ff3a26335237da60ff737de1e008db3b3fc20fd6d3912bf1c"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.649365 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.651248 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" event={"ID":"b18114f1-3f03-4f73-b46e-62056032d351","Type":"ContainerStarted","Data":"9d5f921b4039b68cae28ce476e2fa4a047d1920720bae7ca82cde0f4d59ffe30"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.651406 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.653107 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" event={"ID":"914424ae-a0ec-4578-9a57-824fceadcf0b","Type":"ContainerStarted","Data":"563fb5b3f3a044e248e07b1a20faf1287520c8e18a9f213a64285a24dc451985"} Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.653219 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.660029 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.022092006 podStartE2EDuration="14.660010733s" podCreationTimestamp="2025-09-30 16:58:07 +0000 UTC" firstStartedPulling="2025-09-30 16:58:08.60771029 +0000 UTC m=+2780.620988807" lastFinishedPulling="2025-09-30 16:58:20.245628997 +0000 UTC m=+2792.258907534" observedRunningTime="2025-09-30 16:58:21.656401383 +0000 UTC m=+2793.669679950" watchObservedRunningTime="2025-09-30 16:58:21.660010733 +0000 UTC m=+2793.673289260" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.696426 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=3.7665088620000002 podStartE2EDuration="11.696399628s" podCreationTimestamp="2025-09-30 16:58:10 +0000 UTC" firstStartedPulling="2025-09-30 16:58:12.39926452 +0000 UTC m=+2784.412543047" lastFinishedPulling="2025-09-30 16:58:20.329155286 +0000 UTC m=+2792.342433813" observedRunningTime="2025-09-30 16:58:21.686273766 +0000 UTC m=+2793.699552303" watchObservedRunningTime="2025-09-30 16:58:21.696399628 +0000 UTC m=+2793.709678165" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.712568 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=3.861451218 podStartE2EDuration="11.712548498s" podCreationTimestamp="2025-09-30 16:58:10 +0000 UTC" firstStartedPulling="2025-09-30 16:58:12.469736984 +0000 UTC m=+2784.483015511" lastFinishedPulling="2025-09-30 16:58:20.320834234 +0000 UTC m=+2792.334112791" observedRunningTime="2025-09-30 16:58:21.705533592 +0000 UTC m=+2793.718812139" watchObservedRunningTime="2025-09-30 16:58:21.712548498 +0000 UTC m=+2793.725827045" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.728592 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=4.764232893 podStartE2EDuration="11.728574325s" podCreationTimestamp="2025-09-30 16:58:10 +0000 UTC" firstStartedPulling="2025-09-30 16:58:13.357362746 +0000 UTC m=+2785.370641273" lastFinishedPulling="2025-09-30 16:58:20.321704178 +0000 UTC m=+2792.334982705" observedRunningTime="2025-09-30 16:58:21.72409295 +0000 UTC m=+2793.737371497" watchObservedRunningTime="2025-09-30 16:58:21.728574325 +0000 UTC m=+2793.741852852" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.826801 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" podStartSLOduration=2.984723939 podStartE2EDuration="11.826782822s" podCreationTimestamp="2025-09-30 16:58:10 +0000 UTC" firstStartedPulling="2025-09-30 16:58:11.430430703 +0000 UTC m=+2783.443709230" lastFinishedPulling="2025-09-30 16:58:20.272489586 +0000 UTC m=+2792.285768113" observedRunningTime="2025-09-30 16:58:21.824614322 +0000 UTC m=+2793.837892849" watchObservedRunningTime="2025-09-30 16:58:21.826782822 +0000 UTC m=+2793.840061349" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.832668 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" podStartSLOduration=3.830630715 podStartE2EDuration="12.832641846s" podCreationTimestamp="2025-09-30 16:58:09 +0000 UTC" firstStartedPulling="2025-09-30 16:58:11.245030395 +0000 UTC m=+2783.258308922" lastFinishedPulling="2025-09-30 16:58:20.247041526 +0000 UTC m=+2792.260320053" observedRunningTime="2025-09-30 16:58:21.800217232 +0000 UTC m=+2793.813495759" watchObservedRunningTime="2025-09-30 16:58:21.832641846 +0000 UTC m=+2793.845920383" Sep 30 16:58:21 crc kubenswrapper[4796]: I0930 16:58:21.861930 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" podStartSLOduration=2.758287418 podStartE2EDuration="11.861909751s" podCreationTimestamp="2025-09-30 16:58:10 +0000 UTC" firstStartedPulling="2025-09-30 16:58:11.218247809 +0000 UTC m=+2783.231526336" lastFinishedPulling="2025-09-30 16:58:20.321870142 +0000 UTC m=+2792.335148669" observedRunningTime="2025-09-30 16:58:21.841336548 +0000 UTC m=+2793.854615095" watchObservedRunningTime="2025-09-30 16:58:21.861909751 +0000 UTC m=+2793.875188278" Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.664603 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460" exitCode=2 Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.664748 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460"} Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.665059 4796 scope.go:117] "RemoveContainer" containerID="031e851cc11a95f4d26921d631b298c75748344570bbc5d1e8915f6ae36b9906" Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.665677 4796 scope.go:117] "RemoveContainer" containerID="7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460" Sep 30 16:58:22 crc kubenswrapper[4796]: E0930 16:58:22.666444 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 10s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.670549 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044" exitCode=2 Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.672091 4796 scope.go:117] "RemoveContainer" containerID="f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044" Sep 30 16:58:22 crc kubenswrapper[4796]: E0930 16:58:22.672342 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 10s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.672532 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044"} Sep 30 16:58:22 crc kubenswrapper[4796]: I0930 16:58:22.753955 4796 scope.go:117] "RemoveContainer" containerID="40c72b999a335d12493390eabdd70eb3010b0ce7fc570702bde45f93b44fb01d" Sep 30 16:58:23 crc kubenswrapper[4796]: I0930 16:58:23.684234 4796 scope.go:117] "RemoveContainer" containerID="f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044" Sep 30 16:58:23 crc kubenswrapper[4796]: E0930 16:58:23.685033 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 10s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:23 crc kubenswrapper[4796]: I0930 16:58:23.686023 4796 scope.go:117] "RemoveContainer" containerID="7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460" Sep 30 16:58:23 crc kubenswrapper[4796]: E0930 16:58:23.686439 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 10s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:58:25 crc kubenswrapper[4796]: I0930 16:58:25.796795 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:25 crc kubenswrapper[4796]: I0930 16:58:25.799584 4796 scope.go:117] "RemoveContainer" containerID="f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044" Sep 30 16:58:25 crc kubenswrapper[4796]: E0930 16:58:25.799998 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 10s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.343013 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-fa82-account-create-q9fh4"] Sep 30 16:58:26 crc kubenswrapper[4796]: E0930 16:58:26.343439 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="extract" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.343452 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="extract" Sep 30 16:58:26 crc kubenswrapper[4796]: E0930 16:58:26.343465 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="util" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.343471 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="util" Sep 30 16:58:26 crc kubenswrapper[4796]: E0930 16:58:26.343500 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="pull" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.343507 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="pull" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.343692 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d843fc8-0ae6-4498-97b4-9b40bafd59d0" containerName="extract" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.344453 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.346715 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.358356 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-fa82-account-create-q9fh4"] Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.377098 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.377915 4796 scope.go:117] "RemoveContainer" containerID="7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460" Sep 30 16:58:26 crc kubenswrapper[4796]: E0930 16:58:26.378150 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 10s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.414331 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vbq8\" (UniqueName: \"kubernetes.io/projected/a2d3ff67-b482-4050-baad-14165c0f626a-kube-api-access-8vbq8\") pod \"cloudkitty-fa82-account-create-q9fh4\" (UID: \"a2d3ff67-b482-4050-baad-14165c0f626a\") " pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.517323 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vbq8\" (UniqueName: \"kubernetes.io/projected/a2d3ff67-b482-4050-baad-14165c0f626a-kube-api-access-8vbq8\") pod \"cloudkitty-fa82-account-create-q9fh4\" (UID: \"a2d3ff67-b482-4050-baad-14165c0f626a\") " pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.582298 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vbq8\" (UniqueName: \"kubernetes.io/projected/a2d3ff67-b482-4050-baad-14165c0f626a-kube-api-access-8vbq8\") pod \"cloudkitty-fa82-account-create-q9fh4\" (UID: \"a2d3ff67-b482-4050-baad-14165c0f626a\") " pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:26 crc kubenswrapper[4796]: I0930 16:58:26.691473 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:27 crc kubenswrapper[4796]: I0930 16:58:27.519523 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-fa82-account-create-q9fh4"] Sep 30 16:58:27 crc kubenswrapper[4796]: I0930 16:58:27.734532 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-fa82-account-create-q9fh4" event={"ID":"a2d3ff67-b482-4050-baad-14165c0f626a","Type":"ContainerStarted","Data":"008439c86aecc2d79a5db6552342f8f5b6ad1271b7c28f055955a32fb31dc69c"} Sep 30 16:58:28 crc kubenswrapper[4796]: I0930 16:58:28.748858 4796 generic.go:334] "Generic (PLEG): container finished" podID="a2d3ff67-b482-4050-baad-14165c0f626a" containerID="7b1109d8d983c903ae7ec6329b07de960ee50eeb7ff25ee014aa5642c725dc35" exitCode=0 Sep 30 16:58:28 crc kubenswrapper[4796]: I0930 16:58:28.749632 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-fa82-account-create-q9fh4" event={"ID":"a2d3ff67-b482-4050-baad-14165c0f626a","Type":"ContainerDied","Data":"7b1109d8d983c903ae7ec6329b07de960ee50eeb7ff25ee014aa5642c725dc35"} Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.364658 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.525253 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vbq8\" (UniqueName: \"kubernetes.io/projected/a2d3ff67-b482-4050-baad-14165c0f626a-kube-api-access-8vbq8\") pod \"a2d3ff67-b482-4050-baad-14165c0f626a\" (UID: \"a2d3ff67-b482-4050-baad-14165c0f626a\") " Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.531409 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2d3ff67-b482-4050-baad-14165c0f626a-kube-api-access-8vbq8" (OuterVolumeSpecName: "kube-api-access-8vbq8") pod "a2d3ff67-b482-4050-baad-14165c0f626a" (UID: "a2d3ff67-b482-4050-baad-14165c0f626a"). InnerVolumeSpecName "kube-api-access-8vbq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.627285 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vbq8\" (UniqueName: \"kubernetes.io/projected/a2d3ff67-b482-4050-baad-14165c0f626a-kube-api-access-8vbq8\") on node \"crc\" DevicePath \"\"" Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.778111 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-fa82-account-create-q9fh4" event={"ID":"a2d3ff67-b482-4050-baad-14165c0f626a","Type":"ContainerDied","Data":"008439c86aecc2d79a5db6552342f8f5b6ad1271b7c28f055955a32fb31dc69c"} Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.778153 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="008439c86aecc2d79a5db6552342f8f5b6ad1271b7c28f055955a32fb31dc69c" Sep 30 16:58:30 crc kubenswrapper[4796]: I0930 16:58:30.778207 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-fa82-account-create-q9fh4" Sep 30 16:58:31 crc kubenswrapper[4796]: I0930 16:58:31.097620 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:58:31 crc kubenswrapper[4796]: I0930 16:58:31.098144 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.054749 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx"] Sep 30 16:58:32 crc kubenswrapper[4796]: E0930 16:58:32.055267 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2d3ff67-b482-4050-baad-14165c0f626a" containerName="mariadb-account-create" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.055285 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2d3ff67-b482-4050-baad-14165c0f626a" containerName="mariadb-account-create" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.055506 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2d3ff67-b482-4050-baad-14165c0f626a" containerName="mariadb-account-create" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.056279 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.058573 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.058612 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-bbxns" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.058766 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.082115 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.162269 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8kf6\" (UniqueName: \"kubernetes.io/projected/c18ee88c-1de9-4356-95bd-6df7da966490-kube-api-access-t8kf6\") pod \"obo-prometheus-operator-7c8cf85677-l7klx\" (UID: \"c18ee88c-1de9-4356-95bd-6df7da966490\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.215326 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.216660 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.218964 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.220327 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-b69sx" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.233539 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.234911 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.246587 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.264397 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8kf6\" (UniqueName: \"kubernetes.io/projected/c18ee88c-1de9-4356-95bd-6df7da966490-kube-api-access-t8kf6\") pod \"obo-prometheus-operator-7c8cf85677-l7klx\" (UID: \"c18ee88c-1de9-4356-95bd-6df7da966490\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.283617 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8kf6\" (UniqueName: \"kubernetes.io/projected/c18ee88c-1de9-4356-95bd-6df7da966490-kube-api-access-t8kf6\") pod \"obo-prometheus-operator-7c8cf85677-l7klx\" (UID: \"c18ee88c-1de9-4356-95bd-6df7da966490\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.305049 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.366325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a06303-d13f-4a67-b856-95e577e0fab7-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z\" (UID: \"a9a06303-d13f-4a67-b856-95e577e0fab7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.366646 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1277eb0-bc20-4e0e-a3f8-7a25cbf30168-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6\" (UID: \"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.366699 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1277eb0-bc20-4e0e-a3f8-7a25cbf30168-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6\" (UID: \"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.366836 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a06303-d13f-4a67-b856-95e577e0fab7-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z\" (UID: \"a9a06303-d13f-4a67-b856-95e577e0fab7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.403905 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-929vt"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.405675 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.409055 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-nmjxl" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.409265 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.414814 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.447750 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-929vt"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.470421 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1277eb0-bc20-4e0e-a3f8-7a25cbf30168-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6\" (UID: \"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.470505 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a06303-d13f-4a67-b856-95e577e0fab7-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z\" (UID: \"a9a06303-d13f-4a67-b856-95e577e0fab7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.470563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gttj8\" (UniqueName: \"kubernetes.io/projected/0454eedd-98d0-449a-ab54-ceb594db9c5e-kube-api-access-gttj8\") pod \"observability-operator-cc5f78dfc-929vt\" (UID: \"0454eedd-98d0-449a-ab54-ceb594db9c5e\") " pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.470642 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a06303-d13f-4a67-b856-95e577e0fab7-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z\" (UID: \"a9a06303-d13f-4a67-b856-95e577e0fab7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.470674 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0454eedd-98d0-449a-ab54-ceb594db9c5e-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-929vt\" (UID: \"0454eedd-98d0-449a-ab54-ceb594db9c5e\") " pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.470698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1277eb0-bc20-4e0e-a3f8-7a25cbf30168-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6\" (UID: \"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.487896 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a06303-d13f-4a67-b856-95e577e0fab7-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z\" (UID: \"a9a06303-d13f-4a67-b856-95e577e0fab7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.487929 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a06303-d13f-4a67-b856-95e577e0fab7-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z\" (UID: \"a9a06303-d13f-4a67-b856-95e577e0fab7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.488356 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1277eb0-bc20-4e0e-a3f8-7a25cbf30168-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6\" (UID: \"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.506189 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1277eb0-bc20-4e0e-a3f8-7a25cbf30168-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6\" (UID: \"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.565534 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.570768 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.572069 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0454eedd-98d0-449a-ab54-ceb594db9c5e-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-929vt\" (UID: \"0454eedd-98d0-449a-ab54-ceb594db9c5e\") " pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.572175 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gttj8\" (UniqueName: \"kubernetes.io/projected/0454eedd-98d0-449a-ab54-ceb594db9c5e-kube-api-access-gttj8\") pod \"observability-operator-cc5f78dfc-929vt\" (UID: \"0454eedd-98d0-449a-ab54-ceb594db9c5e\") " pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.585593 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0454eedd-98d0-449a-ab54-ceb594db9c5e-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-929vt\" (UID: \"0454eedd-98d0-449a-ab54-ceb594db9c5e\") " pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.599749 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gttj8\" (UniqueName: \"kubernetes.io/projected/0454eedd-98d0-449a-ab54-ceb594db9c5e-kube-api-access-gttj8\") pod \"observability-operator-cc5f78dfc-929vt\" (UID: \"0454eedd-98d0-449a-ab54-ceb594db9c5e\") " pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.609104 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-xvbbv"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.610468 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.615097 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-9nz8k" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.619750 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-xvbbv"] Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.673830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e62a431-bc1c-4624-8463-93a4f9b7b32e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-xvbbv\" (UID: \"1e62a431-bc1c-4624-8463-93a4f9b7b32e\") " pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.674018 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj4cl\" (UniqueName: \"kubernetes.io/projected/1e62a431-bc1c-4624-8463-93a4f9b7b32e-kube-api-access-tj4cl\") pod \"perses-operator-54bc95c9fb-xvbbv\" (UID: \"1e62a431-bc1c-4624-8463-93a4f9b7b32e\") " pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.740435 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.775323 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e62a431-bc1c-4624-8463-93a4f9b7b32e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-xvbbv\" (UID: \"1e62a431-bc1c-4624-8463-93a4f9b7b32e\") " pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.775491 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj4cl\" (UniqueName: \"kubernetes.io/projected/1e62a431-bc1c-4624-8463-93a4f9b7b32e-kube-api-access-tj4cl\") pod \"perses-operator-54bc95c9fb-xvbbv\" (UID: \"1e62a431-bc1c-4624-8463-93a4f9b7b32e\") " pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.776701 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e62a431-bc1c-4624-8463-93a4f9b7b32e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-xvbbv\" (UID: \"1e62a431-bc1c-4624-8463-93a4f9b7b32e\") " pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:32 crc kubenswrapper[4796]: I0930 16:58:32.820033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj4cl\" (UniqueName: \"kubernetes.io/projected/1e62a431-bc1c-4624-8463-93a4f9b7b32e-kube-api-access-tj4cl\") pod \"perses-operator-54bc95c9fb-xvbbv\" (UID: \"1e62a431-bc1c-4624-8463-93a4f9b7b32e\") " pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.031236 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx"] Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.031535 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.343299 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6"] Sep 30 16:58:33 crc kubenswrapper[4796]: W0930 16:58:33.357844 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1277eb0_bc20_4e0e_a3f8_7a25cbf30168.slice/crio-4fbc2df9b845b60ceed56e0ee59ead2d7551ee65f1c1a921574e39fe2638e9c5 WatchSource:0}: Error finding container 4fbc2df9b845b60ceed56e0ee59ead2d7551ee65f1c1a921574e39fe2638e9c5: Status 404 returned error can't find the container with id 4fbc2df9b845b60ceed56e0ee59ead2d7551ee65f1c1a921574e39fe2638e9c5 Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.441929 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z"] Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.591307 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-929vt"] Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.787100 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-xvbbv"] Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.815769 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" event={"ID":"0454eedd-98d0-449a-ab54-ceb594db9c5e","Type":"ContainerStarted","Data":"4a7a39c5c0e3ce9193600acf6c33d13c756ad9e01cfd4022904ae74051a1f4da"} Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.817475 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" event={"ID":"c18ee88c-1de9-4356-95bd-6df7da966490","Type":"ContainerStarted","Data":"9e9cbeb96be9fb9c3d446da329296555bc0c2715ae827f7d000b287f5c9cc036"} Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.820939 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" event={"ID":"1e62a431-bc1c-4624-8463-93a4f9b7b32e","Type":"ContainerStarted","Data":"331ebf3bb018de28d73198968186794dd81c90b12efefa53b455d1095196b25c"} Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.821999 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" event={"ID":"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168","Type":"ContainerStarted","Data":"4fbc2df9b845b60ceed56e0ee59ead2d7551ee65f1c1a921574e39fe2638e9c5"} Sep 30 16:58:33 crc kubenswrapper[4796]: I0930 16:58:33.822943 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" event={"ID":"a9a06303-d13f-4a67-b856-95e577e0fab7","Type":"ContainerStarted","Data":"a1a99394e7e8afd3c22c7ca02052846041a4efd38982085ac00b2c842356173d"} Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.892538 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-glbqt"] Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.894635 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.897563 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-mbb9q" Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.898068 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.899708 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.899908 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Sep 30 16:58:36 crc kubenswrapper[4796]: I0930 16:58:36.915506 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-glbqt"] Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.005258 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-config-data\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.005325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-certs\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.005376 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-scripts\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.005426 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqk99\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-kube-api-access-bqk99\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.005450 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-combined-ca-bundle\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.107541 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqk99\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-kube-api-access-bqk99\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.107596 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-combined-ca-bundle\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.107676 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-config-data\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.107711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-certs\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.107762 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-scripts\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.121452 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-config-data\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.123203 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-combined-ca-bundle\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.123608 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-scripts\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.123693 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-certs\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.137068 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqk99\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-kube-api-access-bqk99\") pod \"cloudkitty-db-sync-glbqt\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.231916 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.779105 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-glbqt"] Sep 30 16:58:37 crc kubenswrapper[4796]: I0930 16:58:37.920895 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-glbqt" event={"ID":"8fd89bba-353c-4e91-8821-1ed2d35e2507","Type":"ContainerStarted","Data":"fec4bf3a9dac05e8f841179d2f373ce036458412a6d6a334283497624d7ccfb3"} Sep 30 16:58:38 crc kubenswrapper[4796]: I0930 16:58:38.064962 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 16:58:38 crc kubenswrapper[4796]: I0930 16:58:38.745191 4796 scope.go:117] "RemoveContainer" containerID="f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044" Sep 30 16:58:38 crc kubenswrapper[4796]: I0930 16:58:38.747172 4796 scope.go:117] "RemoveContainer" containerID="7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460" Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.255825 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-bccccd5f6-j4nk9" Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.545370 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-6b6cdc96db-2rjz6" Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.626617 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f" Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.804347 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.997335 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b" exitCode=2 Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.997400 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b"} Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.997436 4796 scope.go:117] "RemoveContainer" containerID="7588069e72e90c47a88a3bfdfd64ad72f1be39abf4584aa2ffaa1db7f7fff460" Sep 30 16:58:40 crc kubenswrapper[4796]: I0930 16:58:40.998155 4796 scope.go:117] "RemoveContainer" containerID="724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b" Sep 30 16:58:40 crc kubenswrapper[4796]: E0930 16:58:40.998553 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.005697 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" exitCode=2 Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.005741 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1"} Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.006567 4796 scope.go:117] "RemoveContainer" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" Sep 30 16:58:41 crc kubenswrapper[4796]: E0930 16:58:41.006944 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.376180 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.376512 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.454383 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="6cbd3fc7-5956-4320-b074-bc791068c9cc" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 30 16:58:41 crc kubenswrapper[4796]: I0930 16:58:41.747180 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Sep 30 16:58:42 crc kubenswrapper[4796]: I0930 16:58:42.018289 4796 scope.go:117] "RemoveContainer" containerID="724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b" Sep 30 16:58:42 crc kubenswrapper[4796]: E0930 16:58:42.018546 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:58:42 crc kubenswrapper[4796]: I0930 16:58:42.018993 4796 scope.go:117] "RemoveContainer" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" Sep 30 16:58:42 crc kubenswrapper[4796]: E0930 16:58:42.019182 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:42 crc kubenswrapper[4796]: I0930 16:58:42.841057 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Sep 30 16:58:45 crc kubenswrapper[4796]: I0930 16:58:45.796821 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:58:45 crc kubenswrapper[4796]: I0930 16:58:45.797947 4796 scope.go:117] "RemoveContainer" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" Sep 30 16:58:45 crc kubenswrapper[4796]: E0930 16:58:45.798269 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:51 crc kubenswrapper[4796]: I0930 16:58:51.449465 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="6cbd3fc7-5956-4320-b074-bc791068c9cc" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 30 16:58:54 crc kubenswrapper[4796]: I0930 16:58:54.736608 4796 scope.go:117] "RemoveContainer" containerID="724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b" Sep 30 16:58:54 crc kubenswrapper[4796]: E0930 16:58:54.737248 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:58:56 crc kubenswrapper[4796]: I0930 16:58:56.733405 4796 scope.go:117] "RemoveContainer" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" Sep 30 16:58:56 crc kubenswrapper[4796]: E0930 16:58:56.734023 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:58:57 crc kubenswrapper[4796]: E0930 16:58:57.499806 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e" Sep 30 16:58:57 crc kubenswrapper[4796]: E0930 16:58:57.500219 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e,Command:[],Args:[--namespace=$(NAMESPACE) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=perses=$(RELATED_IMAGE_PERSES) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gttj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-cc5f78dfc-929vt_openshift-operators(0454eedd-98d0-449a-ab54-ceb594db9c5e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:58:57 crc kubenswrapper[4796]: E0930 16:58:57.501808 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" podUID="0454eedd-98d0-449a-ab54-ceb594db9c5e" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.065324 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.065714 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t8kf6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-7c8cf85677-l7klx_openshift-operators(c18ee88c-1de9-4356-95bd-6df7da966490): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.066868 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" podUID="c18ee88c-1de9-4356-95bd-6df7da966490" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.250290 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5\\\"\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" podUID="c18ee88c-1de9-4356-95bd-6df7da966490" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.250723 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e\\\"\"" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" podUID="0454eedd-98d0-449a-ab54-ceb594db9c5e" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.501089 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.504514 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tj4cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-54bc95c9fb-xvbbv_openshift-operators(1e62a431-bc1c-4624-8463-93a4f9b7b32e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 16:58:58 crc kubenswrapper[4796]: E0930 16:58:58.505727 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" podUID="1e62a431-bc1c-4624-8463-93a4f9b7b32e" Sep 30 16:58:59 crc kubenswrapper[4796]: E0930 16:58:59.262728 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c\\\"\"" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" podUID="1e62a431-bc1c-4624-8463-93a4f9b7b32e" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.520247 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rk8bx/must-gather-xkh7b"] Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.522714 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.526109 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rk8bx"/"openshift-service-ca.crt" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.526109 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rk8bx"/"kube-root-ca.crt" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.552170 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rk8bx/must-gather-xkh7b"] Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.566438 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-must-gather-output\") pod \"must-gather-xkh7b\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.566571 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fnsb\" (UniqueName: \"kubernetes.io/projected/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-kube-api-access-9fnsb\") pod \"must-gather-xkh7b\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.668472 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fnsb\" (UniqueName: \"kubernetes.io/projected/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-kube-api-access-9fnsb\") pod \"must-gather-xkh7b\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.669330 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-must-gather-output\") pod \"must-gather-xkh7b\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.670314 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-must-gather-output\") pod \"must-gather-xkh7b\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.690746 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fnsb\" (UniqueName: \"kubernetes.io/projected/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-kube-api-access-9fnsb\") pod \"must-gather-xkh7b\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:00 crc kubenswrapper[4796]: I0930 16:59:00.860487 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.096845 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.096910 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.096967 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.098205 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.098278 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" gracePeriod=600 Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.246742 4796 scope.go:117] "RemoveContainer" containerID="f90b1bdd675bffdb0418cdd3df1fe8307af39f721c5ee2fc0599471620779044" Sep 30 16:59:01 crc kubenswrapper[4796]: E0930 16:59:01.266423 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.287201 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" exitCode=0 Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.287345 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708"} Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.288006 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 16:59:01 crc kubenswrapper[4796]: E0930 16:59:01.288335 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:59:01 crc kubenswrapper[4796]: I0930 16:59:01.448110 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="6cbd3fc7-5956-4320-b074-bc791068c9cc" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 30 16:59:04 crc kubenswrapper[4796]: I0930 16:59:04.794615 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rk8bx/must-gather-xkh7b"] Sep 30 16:59:04 crc kubenswrapper[4796]: I0930 16:59:04.940691 4796 scope.go:117] "RemoveContainer" containerID="aa54084bf09f071e4795b1854e506cfb9ca943c018b90c5bd451399efd8debd9" Sep 30 16:59:04 crc kubenswrapper[4796]: W0930 16:59:04.946899 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90a2b4ab_61f5_49b9_9be8_aeab05d1ba58.slice/crio-c4e1fcca7fe06dad3cd90cec09e76114d0989a3fdf333e18c0bd45e6fe073603 WatchSource:0}: Error finding container c4e1fcca7fe06dad3cd90cec09e76114d0989a3fdf333e18c0bd45e6fe073603: Status 404 returned error can't find the container with id c4e1fcca7fe06dad3cd90cec09e76114d0989a3fdf333e18c0bd45e6fe073603 Sep 30 16:59:04 crc kubenswrapper[4796]: E0930 16:59:04.975677 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/jwysogla/cloudkitty-api@sha256:5541d1160f777174a00982fde3c26a9b32ba156f9f140c9628f66d0eef834c86" Sep 30 16:59:04 crc kubenswrapper[4796]: E0930 16:59:04.975838 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.io/jwysogla/cloudkitty-api@sha256:5541d1160f777174a00982fde3c26a9b32ba156f9f140c9628f66d0eef834c86,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqk99,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-glbqt_openstack(8fd89bba-353c-4e91-8821-1ed2d35e2507): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 16:59:04 crc kubenswrapper[4796]: E0930 16:59:04.977215 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-glbqt" podUID="8fd89bba-353c-4e91-8821-1ed2d35e2507" Sep 30 16:59:05 crc kubenswrapper[4796]: I0930 16:59:05.336320 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" event={"ID":"a9a06303-d13f-4a67-b856-95e577e0fab7","Type":"ContainerStarted","Data":"6a5557fbf942411c6466941325a0b223a06f33d801e3d311112a318001a8aa5a"} Sep 30 16:59:05 crc kubenswrapper[4796]: I0930 16:59:05.338426 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" event={"ID":"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58","Type":"ContainerStarted","Data":"c4e1fcca7fe06dad3cd90cec09e76114d0989a3fdf333e18c0bd45e6fe073603"} Sep 30 16:59:05 crc kubenswrapper[4796]: I0930 16:59:05.345106 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" event={"ID":"f1277eb0-bc20-4e0e-a3f8-7a25cbf30168","Type":"ContainerStarted","Data":"37409857b52e97e1f8d52edbf60625f6e41acca37f28c8acc2b193e8a530f270"} Sep 30 16:59:05 crc kubenswrapper[4796]: E0930 16:59:05.347407 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/jwysogla/cloudkitty-api@sha256:5541d1160f777174a00982fde3c26a9b32ba156f9f140c9628f66d0eef834c86\\\"\"" pod="openstack/cloudkitty-db-sync-glbqt" podUID="8fd89bba-353c-4e91-8821-1ed2d35e2507" Sep 30 16:59:05 crc kubenswrapper[4796]: I0930 16:59:05.371248 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z" podStartSLOduration=5.591473646 podStartE2EDuration="33.371224746s" podCreationTimestamp="2025-09-30 16:58:32 +0000 UTC" firstStartedPulling="2025-09-30 16:58:33.483432162 +0000 UTC m=+2805.496710689" lastFinishedPulling="2025-09-30 16:59:01.263183262 +0000 UTC m=+2833.276461789" observedRunningTime="2025-09-30 16:59:05.359541767 +0000 UTC m=+2837.372820304" watchObservedRunningTime="2025-09-30 16:59:05.371224746 +0000 UTC m=+2837.384503283" Sep 30 16:59:05 crc kubenswrapper[4796]: I0930 16:59:05.431139 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6" podStartSLOduration=7.992996222 podStartE2EDuration="33.431121212s" podCreationTimestamp="2025-09-30 16:58:32 +0000 UTC" firstStartedPulling="2025-09-30 16:58:33.368760186 +0000 UTC m=+2805.382038703" lastFinishedPulling="2025-09-30 16:58:58.806885166 +0000 UTC m=+2830.820163693" observedRunningTime="2025-09-30 16:59:05.412641287 +0000 UTC m=+2837.425919824" watchObservedRunningTime="2025-09-30 16:59:05.431121212 +0000 UTC m=+2837.444399739" Sep 30 16:59:07 crc kubenswrapper[4796]: I0930 16:59:07.734456 4796 scope.go:117] "RemoveContainer" containerID="724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b" Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.400319 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" event={"ID":"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58","Type":"ContainerStarted","Data":"b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe"} Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.401087 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" event={"ID":"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58","Type":"ContainerStarted","Data":"c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35"} Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.403642 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" exitCode=2 Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.403671 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873"} Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.403692 4796 scope.go:117] "RemoveContainer" containerID="724cfc855b28830f47a2becba20a7cc3c25bd117f4b31a7eb1b8992ce512954b" Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.404137 4796 scope.go:117] "RemoveContainer" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" Sep 30 16:59:10 crc kubenswrapper[4796]: E0930 16:59:10.404343 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.420532 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" podStartSLOduration=5.527717572 podStartE2EDuration="10.420513011s" podCreationTimestamp="2025-09-30 16:59:00 +0000 UTC" firstStartedPulling="2025-09-30 16:59:04.960497131 +0000 UTC m=+2836.973775658" lastFinishedPulling="2025-09-30 16:59:09.85329256 +0000 UTC m=+2841.866571097" observedRunningTime="2025-09-30 16:59:10.415770014 +0000 UTC m=+2842.429048541" watchObservedRunningTime="2025-09-30 16:59:10.420513011 +0000 UTC m=+2842.433791538" Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.733570 4796 scope.go:117] "RemoveContainer" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" Sep 30 16:59:10 crc kubenswrapper[4796]: I0930 16:59:10.797315 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.375100 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.375429 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.419053 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" exitCode=2 Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.419145 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea"} Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.419212 4796 scope.go:117] "RemoveContainer" containerID="43d1f8f6551a8af16f79ac2dcb59656a7538182e11c9992d5d93c4be797100a1" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.420403 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:11 crc kubenswrapper[4796]: E0930 16:59:11.421076 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.433535 4796 scope.go:117] "RemoveContainer" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" Sep 30 16:59:11 crc kubenswrapper[4796]: E0930 16:59:11.433809 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.443378 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" event={"ID":"1e62a431-bc1c-4624-8463-93a4f9b7b32e","Type":"ContainerStarted","Data":"07dc045bf9b8926ab3e39d119b70ecf89a9b2efd5b235d047c6098a1e1cf75cb"} Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.445304 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.453303 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Sep 30 16:59:11 crc kubenswrapper[4796]: I0930 16:59:11.549339 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" podStartSLOduration=2.700898315 podStartE2EDuration="39.549319151s" podCreationTimestamp="2025-09-30 16:58:32 +0000 UTC" firstStartedPulling="2025-09-30 16:58:33.79651266 +0000 UTC m=+2805.809791187" lastFinishedPulling="2025-09-30 16:59:10.644933486 +0000 UTC m=+2842.658212023" observedRunningTime="2025-09-30 16:59:11.539066214 +0000 UTC m=+2843.552344751" watchObservedRunningTime="2025-09-30 16:59:11.549319151 +0000 UTC m=+2843.562597678" Sep 30 16:59:12 crc kubenswrapper[4796]: I0930 16:59:12.462303 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:12 crc kubenswrapper[4796]: E0930 16:59:12.463037 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:13 crc kubenswrapper[4796]: I0930 16:59:13.472777 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" event={"ID":"0454eedd-98d0-449a-ab54-ceb594db9c5e","Type":"ContainerStarted","Data":"b7a1e804904863da8b7396c1d081c842c8e3d0611b6274a8aede8d6a1bb96ca4"} Sep 30 16:59:13 crc kubenswrapper[4796]: I0930 16:59:13.473553 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:59:13 crc kubenswrapper[4796]: I0930 16:59:13.503112 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" podStartSLOduration=2.378011307 podStartE2EDuration="41.503093141s" podCreationTimestamp="2025-09-30 16:58:32 +0000 UTC" firstStartedPulling="2025-09-30 16:58:33.60425641 +0000 UTC m=+2805.617534937" lastFinishedPulling="2025-09-30 16:59:12.729338244 +0000 UTC m=+2844.742616771" observedRunningTime="2025-09-30 16:59:13.500485856 +0000 UTC m=+2845.513764383" watchObservedRunningTime="2025-09-30 16:59:13.503093141 +0000 UTC m=+2845.516371658" Sep 30 16:59:13 crc kubenswrapper[4796]: I0930 16:59:13.519689 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-929vt" Sep 30 16:59:13 crc kubenswrapper[4796]: I0930 16:59:13.733866 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 16:59:13 crc kubenswrapper[4796]: E0930 16:59:13.734206 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:59:15 crc kubenswrapper[4796]: I0930 16:59:15.797405 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:59:15 crc kubenswrapper[4796]: I0930 16:59:15.798724 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:15 crc kubenswrapper[4796]: E0930 16:59:15.799015 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:22 crc kubenswrapper[4796]: I0930 16:59:22.575849 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" event={"ID":"c18ee88c-1de9-4356-95bd-6df7da966490","Type":"ContainerStarted","Data":"34f92d27a43adb89fcbf49f8f2eb6cb834b3d733bd463507521f5e06ea7193b2"} Sep 30 16:59:22 crc kubenswrapper[4796]: I0930 16:59:22.601702 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-l7klx" podStartSLOduration=2.319643258 podStartE2EDuration="50.601685479s" podCreationTimestamp="2025-09-30 16:58:32 +0000 UTC" firstStartedPulling="2025-09-30 16:58:33.10738405 +0000 UTC m=+2805.120662577" lastFinishedPulling="2025-09-30 16:59:21.389426261 +0000 UTC m=+2853.402704798" observedRunningTime="2025-09-30 16:59:22.601566255 +0000 UTC m=+2854.614844792" watchObservedRunningTime="2025-09-30 16:59:22.601685479 +0000 UTC m=+2854.614964006" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.035370 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-xvbbv" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.627127 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-sc8t9"] Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.628881 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.632329 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rk8bx"/"default-dockercfg-zd74m" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.787885 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599fe9b7-954d-408e-b7b8-aa941ac23480-host\") pod \"crc-debug-sc8t9\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.788247 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n658r\" (UniqueName: \"kubernetes.io/projected/599fe9b7-954d-408e-b7b8-aa941ac23480-kube-api-access-n658r\") pod \"crc-debug-sc8t9\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.889963 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599fe9b7-954d-408e-b7b8-aa941ac23480-host\") pod \"crc-debug-sc8t9\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.890038 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n658r\" (UniqueName: \"kubernetes.io/projected/599fe9b7-954d-408e-b7b8-aa941ac23480-kube-api-access-n658r\") pod \"crc-debug-sc8t9\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.890089 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599fe9b7-954d-408e-b7b8-aa941ac23480-host\") pod \"crc-debug-sc8t9\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.923964 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n658r\" (UniqueName: \"kubernetes.io/projected/599fe9b7-954d-408e-b7b8-aa941ac23480-kube-api-access-n658r\") pod \"crc-debug-sc8t9\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: I0930 16:59:23.944055 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 16:59:23 crc kubenswrapper[4796]: W0930 16:59:23.985915 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod599fe9b7_954d_408e_b7b8_aa941ac23480.slice/crio-02f191c4aaf1f4c3a025bc83c33338ab693d717891ce57a36c798bbf4c12a1ee WatchSource:0}: Error finding container 02f191c4aaf1f4c3a025bc83c33338ab693d717891ce57a36c798bbf4c12a1ee: Status 404 returned error can't find the container with id 02f191c4aaf1f4c3a025bc83c33338ab693d717891ce57a36c798bbf4c12a1ee Sep 30 16:59:24 crc kubenswrapper[4796]: I0930 16:59:24.628408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-glbqt" event={"ID":"8fd89bba-353c-4e91-8821-1ed2d35e2507","Type":"ContainerStarted","Data":"59ba8be529747d7b597d0da8394c6e049c0e0956061754a07a23ccb2fa4aff54"} Sep 30 16:59:24 crc kubenswrapper[4796]: I0930 16:59:24.638848 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" event={"ID":"599fe9b7-954d-408e-b7b8-aa941ac23480","Type":"ContainerStarted","Data":"02f191c4aaf1f4c3a025bc83c33338ab693d717891ce57a36c798bbf4c12a1ee"} Sep 30 16:59:24 crc kubenswrapper[4796]: I0930 16:59:24.679237 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-glbqt" podStartSLOduration=3.220288311 podStartE2EDuration="48.679216127s" podCreationTimestamp="2025-09-30 16:58:36 +0000 UTC" firstStartedPulling="2025-09-30 16:58:37.793164146 +0000 UTC m=+2809.806442673" lastFinishedPulling="2025-09-30 16:59:23.252091962 +0000 UTC m=+2855.265370489" observedRunningTime="2025-09-30 16:59:24.671649518 +0000 UTC m=+2856.684928065" watchObservedRunningTime="2025-09-30 16:59:24.679216127 +0000 UTC m=+2856.692494644" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.105503 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.108250 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.113898 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.116249 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-dl58j" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.116477 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.116618 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.136866 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.254405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.254453 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.254473 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.254554 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.254664 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.254779 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tx6f\" (UniqueName: \"kubernetes.io/projected/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-kube-api-access-4tx6f\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.371989 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tx6f\" (UniqueName: \"kubernetes.io/projected/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-kube-api-access-4tx6f\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.372151 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.372768 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.372815 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.372865 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.373465 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.373769 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.382782 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.413576 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.417932 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.425711 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tx6f\" (UniqueName: \"kubernetes.io/projected/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-kube-api-access-4tx6f\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.427797 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eefa8387-6e63-41c6-b4c1-c3d6a994dfde-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"eefa8387-6e63-41c6-b4c1-c3d6a994dfde\") " pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.442549 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.733625 4796 scope.go:117] "RemoveContainer" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" Sep 30 16:59:26 crc kubenswrapper[4796]: E0930 16:59:26.734349 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:59:26 crc kubenswrapper[4796]: I0930 16:59:26.737563 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 16:59:26 crc kubenswrapper[4796]: E0930 16:59:26.737805 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.221212 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.248848 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.262390 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.272548 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.272643 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.272827 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.273079 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.273202 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.273460 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-665w5" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.339113 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410365 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410400 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-config\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410422 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410441 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410456 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5526q\" (UniqueName: \"kubernetes.io/projected/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-kube-api-access-5526q\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410487 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410528 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.410612 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512582 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512696 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512716 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-config\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512740 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512757 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512771 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5526q\" (UniqueName: \"kubernetes.io/projected/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-kube-api-access-5526q\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512799 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.512842 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.514033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.527578 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.527615 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6acaf2879198b0f4fa7dfbdab53612edb7b9e7b7209939b626527fd61b8d6871/globalmount\"" pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.528728 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-config\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.531847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.533489 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.533718 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.535649 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.537625 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5526q\" (UniqueName: \"kubernetes.io/projected/65d8e2ee-85e4-4ea5-82ad-5de51cb4e427-kube-api-access-5526q\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.598816 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8a06f85f-900e-4a1a-976e-b2eaf7437ba5\") pod \"prometheus-metric-storage-0\" (UID: \"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427\") " pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.620265 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 16:59:27 crc kubenswrapper[4796]: I0930 16:59:27.737122 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"eefa8387-6e63-41c6-b4c1-c3d6a994dfde","Type":"ContainerStarted","Data":"38c399116a9ca3ba341a57b00a4d4db9217935d5630929f740e2b4c0d3d03d60"} Sep 30 16:59:28 crc kubenswrapper[4796]: I0930 16:59:28.218570 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 16:59:28 crc kubenswrapper[4796]: W0930 16:59:28.225676 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d8e2ee_85e4_4ea5_82ad_5de51cb4e427.slice/crio-b19cfc6dde17df6c3ee811287b53c33595964b5375b9b6ffbb25a941bbb6a37a WatchSource:0}: Error finding container b19cfc6dde17df6c3ee811287b53c33595964b5375b9b6ffbb25a941bbb6a37a: Status 404 returned error can't find the container with id b19cfc6dde17df6c3ee811287b53c33595964b5375b9b6ffbb25a941bbb6a37a Sep 30 16:59:28 crc kubenswrapper[4796]: I0930 16:59:28.745177 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:28 crc kubenswrapper[4796]: E0930 16:59:28.745373 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:28 crc kubenswrapper[4796]: I0930 16:59:28.757452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427","Type":"ContainerStarted","Data":"b19cfc6dde17df6c3ee811287b53c33595964b5375b9b6ffbb25a941bbb6a37a"} Sep 30 16:59:33 crc kubenswrapper[4796]: I0930 16:59:33.816408 4796 generic.go:334] "Generic (PLEG): container finished" podID="8fd89bba-353c-4e91-8821-1ed2d35e2507" containerID="59ba8be529747d7b597d0da8394c6e049c0e0956061754a07a23ccb2fa4aff54" exitCode=0 Sep 30 16:59:33 crc kubenswrapper[4796]: I0930 16:59:33.816503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-glbqt" event={"ID":"8fd89bba-353c-4e91-8821-1ed2d35e2507","Type":"ContainerDied","Data":"59ba8be529747d7b597d0da8394c6e049c0e0956061754a07a23ccb2fa4aff54"} Sep 30 16:59:34 crc kubenswrapper[4796]: I0930 16:59:34.828996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"eefa8387-6e63-41c6-b4c1-c3d6a994dfde","Type":"ContainerStarted","Data":"b6e6fa6e02ca9005d26202a1a72fa9c072ea73689e47253fdf072d4ea86c74c5"} Sep 30 16:59:34 crc kubenswrapper[4796]: I0930 16:59:34.830558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427","Type":"ContainerStarted","Data":"e074678560ed91be21517b094352e72d191a0f83bc76841d0941d060854dfa20"} Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.694702 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.741489 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 16:59:38 crc kubenswrapper[4796]: E0930 16:59:38.741803 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.758326 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-config-data\") pod \"8fd89bba-353c-4e91-8821-1ed2d35e2507\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.758421 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-certs\") pod \"8fd89bba-353c-4e91-8821-1ed2d35e2507\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.758577 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-scripts\") pod \"8fd89bba-353c-4e91-8821-1ed2d35e2507\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.758705 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-combined-ca-bundle\") pod \"8fd89bba-353c-4e91-8821-1ed2d35e2507\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.758877 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqk99\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-kube-api-access-bqk99\") pod \"8fd89bba-353c-4e91-8821-1ed2d35e2507\" (UID: \"8fd89bba-353c-4e91-8821-1ed2d35e2507\") " Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.771848 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-kube-api-access-bqk99" (OuterVolumeSpecName: "kube-api-access-bqk99") pod "8fd89bba-353c-4e91-8821-1ed2d35e2507" (UID: "8fd89bba-353c-4e91-8821-1ed2d35e2507"). InnerVolumeSpecName "kube-api-access-bqk99". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.772470 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-certs" (OuterVolumeSpecName: "certs") pod "8fd89bba-353c-4e91-8821-1ed2d35e2507" (UID: "8fd89bba-353c-4e91-8821-1ed2d35e2507"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.773237 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-scripts" (OuterVolumeSpecName: "scripts") pod "8fd89bba-353c-4e91-8821-1ed2d35e2507" (UID: "8fd89bba-353c-4e91-8821-1ed2d35e2507"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.807191 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fd89bba-353c-4e91-8821-1ed2d35e2507" (UID: "8fd89bba-353c-4e91-8821-1ed2d35e2507"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.818631 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-config-data" (OuterVolumeSpecName: "config-data") pod "8fd89bba-353c-4e91-8821-1ed2d35e2507" (UID: "8fd89bba-353c-4e91-8821-1ed2d35e2507"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.861852 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.861891 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.861907 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqk99\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-kube-api-access-bqk99\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.861919 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd89bba-353c-4e91-8821-1ed2d35e2507-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.861939 4796 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/8fd89bba-353c-4e91-8821-1ed2d35e2507-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.885040 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-glbqt" event={"ID":"8fd89bba-353c-4e91-8821-1ed2d35e2507","Type":"ContainerDied","Data":"fec4bf3a9dac05e8f841179d2f373ce036458412a6d6a334283497624d7ccfb3"} Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.885087 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fec4bf3a9dac05e8f841179d2f373ce036458412a6d6a334283497624d7ccfb3" Sep 30 16:59:38 crc kubenswrapper[4796]: I0930 16:59:38.885104 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-glbqt" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.733594 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:39 crc kubenswrapper[4796]: E0930 16:59:39.734377 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.802012 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-5s8tn"] Sep 30 16:59:39 crc kubenswrapper[4796]: E0930 16:59:39.802493 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd89bba-353c-4e91-8821-1ed2d35e2507" containerName="cloudkitty-db-sync" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.802519 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd89bba-353c-4e91-8821-1ed2d35e2507" containerName="cloudkitty-db-sync" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.802794 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd89bba-353c-4e91-8821-1ed2d35e2507" containerName="cloudkitty-db-sync" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.803903 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.807684 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.808426 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-mbb9q" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.808587 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.808726 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.818056 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-5s8tn"] Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.881592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-combined-ca-bundle\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.881750 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsqzj\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-kube-api-access-dsqzj\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.881785 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-config-data\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.881939 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-certs\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.882004 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-scripts\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.897965 4796 generic.go:334] "Generic (PLEG): container finished" podID="eefa8387-6e63-41c6-b4c1-c3d6a994dfde" containerID="b6e6fa6e02ca9005d26202a1a72fa9c072ea73689e47253fdf072d4ea86c74c5" exitCode=0 Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.898050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"eefa8387-6e63-41c6-b4c1-c3d6a994dfde","Type":"ContainerDied","Data":"b6e6fa6e02ca9005d26202a1a72fa9c072ea73689e47253fdf072d4ea86c74c5"} Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.901720 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" event={"ID":"599fe9b7-954d-408e-b7b8-aa941ac23480","Type":"ContainerStarted","Data":"e21a17e48af95794a5fb3b5abbd4975743723673ffb0e75c19b211ada2d08712"} Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.941540 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" podStartSLOduration=1.905866478 podStartE2EDuration="16.941522983s" podCreationTimestamp="2025-09-30 16:59:23 +0000 UTC" firstStartedPulling="2025-09-30 16:59:23.988132766 +0000 UTC m=+2856.001411293" lastFinishedPulling="2025-09-30 16:59:39.023789251 +0000 UTC m=+2871.037067798" observedRunningTime="2025-09-30 16:59:39.936839167 +0000 UTC m=+2871.950117694" watchObservedRunningTime="2025-09-30 16:59:39.941522983 +0000 UTC m=+2871.954801500" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.983735 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsqzj\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-kube-api-access-dsqzj\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.983794 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-config-data\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.983915 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-certs\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.983947 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-scripts\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.984087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-combined-ca-bundle\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.988853 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-certs\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.989575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-config-data\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.990450 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-combined-ca-bundle\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:39 crc kubenswrapper[4796]: I0930 16:59:39.990647 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-scripts\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:40 crc kubenswrapper[4796]: I0930 16:59:40.001124 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsqzj\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-kube-api-access-dsqzj\") pod \"cloudkitty-storageinit-5s8tn\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:40 crc kubenswrapper[4796]: I0930 16:59:40.127867 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:40 crc kubenswrapper[4796]: I0930 16:59:40.592729 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-5s8tn"] Sep 30 16:59:40 crc kubenswrapper[4796]: W0930 16:59:40.600247 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd96b76a9_dd08_43c3_a7e8_2bc542706345.slice/crio-820438fc0779933d0454b66b1c00c036030011c1c3cf57be7288b8e09b70c001 WatchSource:0}: Error finding container 820438fc0779933d0454b66b1c00c036030011c1c3cf57be7288b8e09b70c001: Status 404 returned error can't find the container with id 820438fc0779933d0454b66b1c00c036030011c1c3cf57be7288b8e09b70c001 Sep 30 16:59:40 crc kubenswrapper[4796]: I0930 16:59:40.796764 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 16:59:40 crc kubenswrapper[4796]: I0930 16:59:40.797478 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:40 crc kubenswrapper[4796]: E0930 16:59:40.797761 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:40 crc kubenswrapper[4796]: I0930 16:59:40.912777 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-5s8tn" event={"ID":"d96b76a9-dd08-43c3-a7e8-2bc542706345","Type":"ContainerStarted","Data":"820438fc0779933d0454b66b1c00c036030011c1c3cf57be7288b8e09b70c001"} Sep 30 16:59:41 crc kubenswrapper[4796]: I0930 16:59:41.735331 4796 scope.go:117] "RemoveContainer" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" Sep 30 16:59:41 crc kubenswrapper[4796]: E0930 16:59:41.736823 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:59:41 crc kubenswrapper[4796]: I0930 16:59:41.933066 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-5s8tn" event={"ID":"d96b76a9-dd08-43c3-a7e8-2bc542706345","Type":"ContainerStarted","Data":"21d7f42438eac9b70b91b6dac13734b94c2566debce339ebde995b2df3719adb"} Sep 30 16:59:41 crc kubenswrapper[4796]: I0930 16:59:41.935374 4796 generic.go:334] "Generic (PLEG): container finished" podID="65d8e2ee-85e4-4ea5-82ad-5de51cb4e427" containerID="e074678560ed91be21517b094352e72d191a0f83bc76841d0941d060854dfa20" exitCode=0 Sep 30 16:59:41 crc kubenswrapper[4796]: I0930 16:59:41.935405 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427","Type":"ContainerDied","Data":"e074678560ed91be21517b094352e72d191a0f83bc76841d0941d060854dfa20"} Sep 30 16:59:41 crc kubenswrapper[4796]: I0930 16:59:41.960178 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-5s8tn" podStartSLOduration=2.960161744 podStartE2EDuration="2.960161744s" podCreationTimestamp="2025-09-30 16:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:59:41.949416843 +0000 UTC m=+2873.962695370" watchObservedRunningTime="2025-09-30 16:59:41.960161744 +0000 UTC m=+2873.973440271" Sep 30 16:59:43 crc kubenswrapper[4796]: I0930 16:59:43.961738 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"eefa8387-6e63-41c6-b4c1-c3d6a994dfde","Type":"ContainerStarted","Data":"af7f9ab9d0e6fbabdee602d2947aa31d44ba9c386b1be28a61d99286fbdb9dd0"} Sep 30 16:59:45 crc kubenswrapper[4796]: I0930 16:59:45.984698 4796 generic.go:334] "Generic (PLEG): container finished" podID="d96b76a9-dd08-43c3-a7e8-2bc542706345" containerID="21d7f42438eac9b70b91b6dac13734b94c2566debce339ebde995b2df3719adb" exitCode=0 Sep 30 16:59:45 crc kubenswrapper[4796]: I0930 16:59:45.985237 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-5s8tn" event={"ID":"d96b76a9-dd08-43c3-a7e8-2bc542706345","Type":"ContainerDied","Data":"21d7f42438eac9b70b91b6dac13734b94c2566debce339ebde995b2df3719adb"} Sep 30 16:59:46 crc kubenswrapper[4796]: I0930 16:59:46.998761 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"eefa8387-6e63-41c6-b4c1-c3d6a994dfde","Type":"ContainerStarted","Data":"c32d22cccad1a71b5f68cc0a51f9aa69c08d60eb1474eca96f29151fcc518087"} Sep 30 16:59:47 crc kubenswrapper[4796]: I0930 16:59:47.021402 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.268237014 podStartE2EDuration="21.021381306s" podCreationTimestamp="2025-09-30 16:59:26 +0000 UTC" firstStartedPulling="2025-09-30 16:59:27.264144153 +0000 UTC m=+2859.277422680" lastFinishedPulling="2025-09-30 16:59:43.017288445 +0000 UTC m=+2875.030566972" observedRunningTime="2025-09-30 16:59:47.021043236 +0000 UTC m=+2879.034321843" watchObservedRunningTime="2025-09-30 16:59:47.021381306 +0000 UTC m=+2879.034659833" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.006774 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.010164 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.816743 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.867467 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsqzj\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-kube-api-access-dsqzj\") pod \"d96b76a9-dd08-43c3-a7e8-2bc542706345\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.867688 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-config-data\") pod \"d96b76a9-dd08-43c3-a7e8-2bc542706345\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.867762 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-scripts\") pod \"d96b76a9-dd08-43c3-a7e8-2bc542706345\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.867792 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-certs\") pod \"d96b76a9-dd08-43c3-a7e8-2bc542706345\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.867812 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-combined-ca-bundle\") pod \"d96b76a9-dd08-43c3-a7e8-2bc542706345\" (UID: \"d96b76a9-dd08-43c3-a7e8-2bc542706345\") " Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.879199 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-scripts" (OuterVolumeSpecName: "scripts") pod "d96b76a9-dd08-43c3-a7e8-2bc542706345" (UID: "d96b76a9-dd08-43c3-a7e8-2bc542706345"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.882436 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-kube-api-access-dsqzj" (OuterVolumeSpecName: "kube-api-access-dsqzj") pod "d96b76a9-dd08-43c3-a7e8-2bc542706345" (UID: "d96b76a9-dd08-43c3-a7e8-2bc542706345"). InnerVolumeSpecName "kube-api-access-dsqzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.890811 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-certs" (OuterVolumeSpecName: "certs") pod "d96b76a9-dd08-43c3-a7e8-2bc542706345" (UID: "d96b76a9-dd08-43c3-a7e8-2bc542706345"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.908796 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-config-data" (OuterVolumeSpecName: "config-data") pod "d96b76a9-dd08-43c3-a7e8-2bc542706345" (UID: "d96b76a9-dd08-43c3-a7e8-2bc542706345"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.912922 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d96b76a9-dd08-43c3-a7e8-2bc542706345" (UID: "d96b76a9-dd08-43c3-a7e8-2bc542706345"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.971120 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.971153 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.971161 4796 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-certs\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.971170 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96b76a9-dd08-43c3-a7e8-2bc542706345-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:48 crc kubenswrapper[4796]: I0930 16:59:48.971183 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsqzj\" (UniqueName: \"kubernetes.io/projected/d96b76a9-dd08-43c3-a7e8-2bc542706345-kube-api-access-dsqzj\") on node \"crc\" DevicePath \"\"" Sep 30 16:59:49 crc kubenswrapper[4796]: I0930 16:59:49.015300 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-5s8tn" event={"ID":"d96b76a9-dd08-43c3-a7e8-2bc542706345","Type":"ContainerDied","Data":"820438fc0779933d0454b66b1c00c036030011c1c3cf57be7288b8e09b70c001"} Sep 30 16:59:49 crc kubenswrapper[4796]: I0930 16:59:49.015324 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-5s8tn" Sep 30 16:59:49 crc kubenswrapper[4796]: I0930 16:59:49.015349 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="820438fc0779933d0454b66b1c00c036030011c1c3cf57be7288b8e09b70c001" Sep 30 16:59:49 crc kubenswrapper[4796]: I0930 16:59:49.998393 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Sep 30 16:59:49 crc kubenswrapper[4796]: E0930 16:59:49.999224 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96b76a9-dd08-43c3-a7e8-2bc542706345" containerName="cloudkitty-storageinit" Sep 30 16:59:49 crc kubenswrapper[4796]: I0930 16:59:49.999248 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96b76a9-dd08-43c3-a7e8-2bc542706345" containerName="cloudkitty-storageinit" Sep 30 16:59:49 crc kubenswrapper[4796]: I0930 16:59:49.999548 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96b76a9-dd08-43c3-a7e8-2bc542706345" containerName="cloudkitty-storageinit" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.001934 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.010907 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.011098 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.011298 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.011505 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.011682 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-mbb9q" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.046711 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.102775 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t7fw\" (UniqueName: \"kubernetes.io/projected/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-kube-api-access-7t7fw\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.102820 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.102859 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-scripts\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.102989 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-config-data\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.103015 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.103076 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-certs\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.207018 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.207087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-scripts\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.207171 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-config-data\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.207193 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.207275 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-certs\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.207317 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t7fw\" (UniqueName: \"kubernetes.io/projected/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-kube-api-access-7t7fw\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.218333 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.218492 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-scripts\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.218790 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-certs\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.220025 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-config-data\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.241599 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.245054 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.262350 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.262609 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t7fw\" (UniqueName: \"kubernetes.io/projected/1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad-kube-api-access-7t7fw\") pod \"cloudkitty-proc-0\" (UID: \"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad\") " pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.267361 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311134 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-certs\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311272 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-logs\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311300 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311333 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-scripts\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311388 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-config-data\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311437 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.311462 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6k6c\" (UniqueName: \"kubernetes.io/projected/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-kube-api-access-j6k6c\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.340051 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.340552 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.414853 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.414893 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6k6c\" (UniqueName: \"kubernetes.io/projected/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-kube-api-access-j6k6c\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.414946 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-certs\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.415041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-logs\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.415061 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.415087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-scripts\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.415129 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-config-data\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.417247 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-logs\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.424965 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-scripts\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.429544 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.430488 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-config-data\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.432610 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.438612 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-certs\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.465530 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6k6c\" (UniqueName: \"kubernetes.io/projected/dd002cf0-64b3-4fae-8624-a7114c4ea6bf-kube-api-access-j6k6c\") pod \"cloudkitty-api-0\" (UID: \"dd002cf0-64b3-4fae-8624-a7114c4ea6bf\") " pod="openstack/cloudkitty-api-0" Sep 30 16:59:50 crc kubenswrapper[4796]: I0930 16:59:50.725448 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Sep 30 16:59:51 crc kubenswrapper[4796]: I0930 16:59:51.733590 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 16:59:51 crc kubenswrapper[4796]: E0930 16:59:51.734064 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 16:59:53 crc kubenswrapper[4796]: I0930 16:59:53.733274 4796 scope.go:117] "RemoveContainer" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" Sep 30 16:59:54 crc kubenswrapper[4796]: I0930 16:59:54.733919 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:55 crc kubenswrapper[4796]: I0930 16:59:55.676213 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Sep 30 16:59:55 crc kubenswrapper[4796]: W0930 16:59:55.679658 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1dd26425_9aaf_4a1a_be8e_1fbfd2d248ad.slice/crio-c4f2e939fa2df2ca90f058da9904cebb544dffaee24c10b6e42ded81012765b3 WatchSource:0}: Error finding container c4f2e939fa2df2ca90f058da9904cebb544dffaee24c10b6e42ded81012765b3: Status 404 returned error can't find the container with id c4f2e939fa2df2ca90f058da9904cebb544dffaee24c10b6e42ded81012765b3 Sep 30 16:59:55 crc kubenswrapper[4796]: I0930 16:59:55.726006 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.102325 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" exitCode=2 Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.102418 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26"} Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.104123 4796 scope.go:117] "RemoveContainer" containerID="c5b632e5e4ab6eb6e07db10559ca4922ae9d549fd3289ee1e819464f7993b873" Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.105354 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"dd002cf0-64b3-4fae-8624-a7114c4ea6bf","Type":"ContainerStarted","Data":"3462f083e505a87e4e15bc454749b785454a48ccbf363dd67efe13f614de1080"} Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.105412 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 16:59:56 crc kubenswrapper[4796]: E0930 16:59:56.105829 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.106963 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"c4f2e939fa2df2ca90f058da9904cebb544dffaee24c10b6e42ded81012765b3"} Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.116946 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" exitCode=2 Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.117046 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53"} Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.117793 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 16:59:56 crc kubenswrapper[4796]: E0930 16:59:56.118088 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.121238 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427","Type":"ContainerStarted","Data":"04ba2131ebb139fe1a2537416e7b3dfc8e24008358407cf6d2c649aa9110acfb"} Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.341396 4796 scope.go:117] "RemoveContainer" containerID="a89bbd5d0cb9c264a6222eae7070907e45f769bd2b00ff94f2c1925a2f04b9ea" Sep 30 16:59:56 crc kubenswrapper[4796]: I0930 16:59:56.375237 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 16:59:57 crc kubenswrapper[4796]: I0930 16:59:57.137594 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 16:59:57 crc kubenswrapper[4796]: E0930 16:59:57.138268 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 16:59:57 crc kubenswrapper[4796]: I0930 16:59:57.139181 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"dd002cf0-64b3-4fae-8624-a7114c4ea6bf","Type":"ContainerStarted","Data":"74d21dd86872235b01f81f7a00ee806ed3452cec27aada85d2201653b690da52"} Sep 30 16:59:57 crc kubenswrapper[4796]: I0930 16:59:57.139211 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"dd002cf0-64b3-4fae-8624-a7114c4ea6bf","Type":"ContainerStarted","Data":"7a97c50852b0fcf5b1b010a1a003c5beed82333150cca7fb28fd88e7a0e42d84"} Sep 30 16:59:57 crc kubenswrapper[4796]: I0930 16:59:57.139312 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Sep 30 16:59:57 crc kubenswrapper[4796]: I0930 16:59:57.171647 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=7.171633385 podStartE2EDuration="7.171633385s" podCreationTimestamp="2025-09-30 16:59:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 16:59:57.169269417 +0000 UTC m=+2889.182547944" watchObservedRunningTime="2025-09-30 16:59:57.171633385 +0000 UTC m=+2889.184911912" Sep 30 16:59:59 crc kubenswrapper[4796]: I0930 16:59:59.162404 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427","Type":"ContainerStarted","Data":"c5cf5049632e84139b06afb2912318c1916b4753ec31feffe06d3597a76e7158"} Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.137625 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2"] Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.139919 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.144287 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.144287 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.149348 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2"] Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.249750 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4052d15-4933-47cb-b44c-eaf792717a77-config-volume\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.249942 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4052d15-4933-47cb-b44c-eaf792717a77-secret-volume\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.250024 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5vvs\" (UniqueName: \"kubernetes.io/projected/d4052d15-4933-47cb-b44c-eaf792717a77-kube-api-access-m5vvs\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.351808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4052d15-4933-47cb-b44c-eaf792717a77-secret-volume\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.351931 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5vvs\" (UniqueName: \"kubernetes.io/projected/d4052d15-4933-47cb-b44c-eaf792717a77-kube-api-access-m5vvs\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.352058 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4052d15-4933-47cb-b44c-eaf792717a77-config-volume\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.353309 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4052d15-4933-47cb-b44c-eaf792717a77-config-volume\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.363747 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4052d15-4933-47cb-b44c-eaf792717a77-secret-volume\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.380897 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5vvs\" (UniqueName: \"kubernetes.io/projected/d4052d15-4933-47cb-b44c-eaf792717a77-kube-api-access-m5vvs\") pod \"collect-profiles-29320860-5v6z2\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.463795 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.796616 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 17:00:00 crc kubenswrapper[4796]: I0930 17:00:00.797679 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:00:00 crc kubenswrapper[4796]: E0930 17:00:00.797931 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:00:01 crc kubenswrapper[4796]: I0930 17:00:01.067436 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2"] Sep 30 17:00:02 crc kubenswrapper[4796]: I0930 17:00:02.195113 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"bfeff6e6601efe13b5d72cab54d22002eb465d32c1d0d581063bb3af5d5bf6ca"} Sep 30 17:00:02 crc kubenswrapper[4796]: I0930 17:00:02.210267 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" event={"ID":"d4052d15-4933-47cb-b44c-eaf792717a77","Type":"ContainerStarted","Data":"87757c26e22722edd2d466f130db34edf97014bf8547192c7b18705b2ad97fac"} Sep 30 17:00:02 crc kubenswrapper[4796]: I0930 17:00:02.210339 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" event={"ID":"d4052d15-4933-47cb-b44c-eaf792717a77","Type":"ContainerStarted","Data":"3e4183470cc7512867a4ba31429087ec951006a4362493d09977ed39c6f960fa"} Sep 30 17:00:02 crc kubenswrapper[4796]: I0930 17:00:02.215367 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=7.37067718 podStartE2EDuration="13.21534808s" podCreationTimestamp="2025-09-30 16:59:49 +0000 UTC" firstStartedPulling="2025-09-30 16:59:55.681657898 +0000 UTC m=+2887.694936425" lastFinishedPulling="2025-09-30 17:00:01.526328798 +0000 UTC m=+2893.539607325" observedRunningTime="2025-09-30 17:00:02.210149229 +0000 UTC m=+2894.223427766" watchObservedRunningTime="2025-09-30 17:00:02.21534808 +0000 UTC m=+2894.228626607" Sep 30 17:00:02 crc kubenswrapper[4796]: I0930 17:00:02.242076 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" podStartSLOduration=2.242052694 podStartE2EDuration="2.242052694s" podCreationTimestamp="2025-09-30 17:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 17:00:02.227844152 +0000 UTC m=+2894.241122679" watchObservedRunningTime="2025-09-30 17:00:02.242052694 +0000 UTC m=+2894.255331221" Sep 30 17:00:03 crc kubenswrapper[4796]: I0930 17:00:03.221721 4796 generic.go:334] "Generic (PLEG): container finished" podID="d4052d15-4933-47cb-b44c-eaf792717a77" containerID="87757c26e22722edd2d466f130db34edf97014bf8547192c7b18705b2ad97fac" exitCode=0 Sep 30 17:00:03 crc kubenswrapper[4796]: I0930 17:00:03.221808 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" event={"ID":"d4052d15-4933-47cb-b44c-eaf792717a77","Type":"ContainerDied","Data":"87757c26e22722edd2d466f130db34edf97014bf8547192c7b18705b2ad97fac"} Sep 30 17:00:05 crc kubenswrapper[4796]: I0930 17:00:05.748592 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 17:00:05 crc kubenswrapper[4796]: I0930 17:00:05.749381 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-central-agent" containerID="cri-o://59877affe9f0dcd0279809ecef684a1b9996fd17cc7da6da6dd7ec8ec67e2e88" gracePeriod=30 Sep 30 17:00:05 crc kubenswrapper[4796]: I0930 17:00:05.749879 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-notification-agent" containerID="cri-o://805681bc59c000c3d10eafcf08305f67ed12c5d256a6d8ffbb4eb0a1a6d7a2ab" gracePeriod=30 Sep 30 17:00:05 crc kubenswrapper[4796]: I0930 17:00:05.749902 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="proxy-httpd" containerID="cri-o://043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159" gracePeriod=30 Sep 30 17:00:05 crc kubenswrapper[4796]: I0930 17:00:05.749944 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="sg-core" containerID="cri-o://42a6b5cb9d6b4b85fe799eedf94866ca97de90de1b559246f517c4346e7e041d" gracePeriod=30 Sep 30 17:00:05 crc kubenswrapper[4796]: I0930 17:00:05.869226 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:05 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:05 crc kubenswrapper[4796]: > Sep 30 17:00:06 crc kubenswrapper[4796]: E0930 17:00:06.232484 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode536dcbd_a48d_409a_8154_49e9e392f3b4.slice/crio-conmon-42a6b5cb9d6b4b85fe799eedf94866ca97de90de1b559246f517c4346e7e041d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode536dcbd_a48d_409a_8154_49e9e392f3b4.slice/crio-043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode536dcbd_a48d_409a_8154_49e9e392f3b4.slice/crio-805681bc59c000c3d10eafcf08305f67ed12c5d256a6d8ffbb4eb0a1a6d7a2ab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode536dcbd_a48d_409a_8154_49e9e392f3b4.slice/crio-conmon-043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159.scope\": RecentStats: unable to find data in memory cache]" Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.261613 4796 generic.go:334] "Generic (PLEG): container finished" podID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerID="043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159" exitCode=0 Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.263516 4796 generic.go:334] "Generic (PLEG): container finished" podID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerID="42a6b5cb9d6b4b85fe799eedf94866ca97de90de1b559246f517c4346e7e041d" exitCode=2 Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.263532 4796 generic.go:334] "Generic (PLEG): container finished" podID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerID="805681bc59c000c3d10eafcf08305f67ed12c5d256a6d8ffbb4eb0a1a6d7a2ab" exitCode=0 Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.261670 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerDied","Data":"043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159"} Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.263577 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerDied","Data":"42a6b5cb9d6b4b85fe799eedf94866ca97de90de1b559246f517c4346e7e041d"} Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.263595 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerDied","Data":"805681bc59c000c3d10eafcf08305f67ed12c5d256a6d8ffbb4eb0a1a6d7a2ab"} Sep 30 17:00:06 crc kubenswrapper[4796]: I0930 17:00:06.733940 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:00:06 crc kubenswrapper[4796]: E0930 17:00:06.734269 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.138323 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.200266 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4052d15-4933-47cb-b44c-eaf792717a77-secret-volume\") pod \"d4052d15-4933-47cb-b44c-eaf792717a77\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.200354 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4052d15-4933-47cb-b44c-eaf792717a77-config-volume\") pod \"d4052d15-4933-47cb-b44c-eaf792717a77\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.200389 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5vvs\" (UniqueName: \"kubernetes.io/projected/d4052d15-4933-47cb-b44c-eaf792717a77-kube-api-access-m5vvs\") pod \"d4052d15-4933-47cb-b44c-eaf792717a77\" (UID: \"d4052d15-4933-47cb-b44c-eaf792717a77\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.202669 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4052d15-4933-47cb-b44c-eaf792717a77-config-volume" (OuterVolumeSpecName: "config-volume") pod "d4052d15-4933-47cb-b44c-eaf792717a77" (UID: "d4052d15-4933-47cb-b44c-eaf792717a77"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.224879 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4052d15-4933-47cb-b44c-eaf792717a77-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d4052d15-4933-47cb-b44c-eaf792717a77" (UID: "d4052d15-4933-47cb-b44c-eaf792717a77"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.231630 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4052d15-4933-47cb-b44c-eaf792717a77-kube-api-access-m5vvs" (OuterVolumeSpecName: "kube-api-access-m5vvs") pod "d4052d15-4933-47cb-b44c-eaf792717a77" (UID: "d4052d15-4933-47cb-b44c-eaf792717a77"). InnerVolumeSpecName "kube-api-access-m5vvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.307626 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4052d15-4933-47cb-b44c-eaf792717a77-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.307942 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4052d15-4933-47cb-b44c-eaf792717a77-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.307957 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5vvs\" (UniqueName: \"kubernetes.io/projected/d4052d15-4933-47cb-b44c-eaf792717a77-kube-api-access-m5vvs\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.330640 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.330678 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320860-5v6z2" event={"ID":"d4052d15-4933-47cb-b44c-eaf792717a77","Type":"ContainerDied","Data":"3e4183470cc7512867a4ba31429087ec951006a4362493d09977ed39c6f960fa"} Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.330715 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e4183470cc7512867a4ba31429087ec951006a4362493d09977ed39c6f960fa" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.343683 4796 generic.go:334] "Generic (PLEG): container finished" podID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerID="59877affe9f0dcd0279809ecef684a1b9996fd17cc7da6da6dd7ec8ec67e2e88" exitCode=0 Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.343746 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerDied","Data":"59877affe9f0dcd0279809ecef684a1b9996fd17cc7da6da6dd7ec8ec67e2e88"} Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.457222 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.512890 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqpz6\" (UniqueName: \"kubernetes.io/projected/e536dcbd-a48d-409a-8154-49e9e392f3b4-kube-api-access-qqpz6\") pod \"e536dcbd-a48d-409a-8154-49e9e392f3b4\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.512992 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-scripts\") pod \"e536dcbd-a48d-409a-8154-49e9e392f3b4\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.513060 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-config-data\") pod \"e536dcbd-a48d-409a-8154-49e9e392f3b4\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.513168 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-sg-core-conf-yaml\") pod \"e536dcbd-a48d-409a-8154-49e9e392f3b4\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.513207 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-run-httpd\") pod \"e536dcbd-a48d-409a-8154-49e9e392f3b4\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.513372 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-log-httpd\") pod \"e536dcbd-a48d-409a-8154-49e9e392f3b4\" (UID: \"e536dcbd-a48d-409a-8154-49e9e392f3b4\") " Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.514685 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e536dcbd-a48d-409a-8154-49e9e392f3b4" (UID: "e536dcbd-a48d-409a-8154-49e9e392f3b4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.521466 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e536dcbd-a48d-409a-8154-49e9e392f3b4" (UID: "e536dcbd-a48d-409a-8154-49e9e392f3b4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.526354 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-scripts" (OuterVolumeSpecName: "scripts") pod "e536dcbd-a48d-409a-8154-49e9e392f3b4" (UID: "e536dcbd-a48d-409a-8154-49e9e392f3b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.535675 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e536dcbd-a48d-409a-8154-49e9e392f3b4-kube-api-access-qqpz6" (OuterVolumeSpecName: "kube-api-access-qqpz6") pod "e536dcbd-a48d-409a-8154-49e9e392f3b4" (UID: "e536dcbd-a48d-409a-8154-49e9e392f3b4"). InnerVolumeSpecName "kube-api-access-qqpz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.577453 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e536dcbd-a48d-409a-8154-49e9e392f3b4" (UID: "e536dcbd-a48d-409a-8154-49e9e392f3b4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.616744 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.616789 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqpz6\" (UniqueName: \"kubernetes.io/projected/e536dcbd-a48d-409a-8154-49e9e392f3b4-kube-api-access-qqpz6\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.616805 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.616816 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.616828 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e536dcbd-a48d-409a-8154-49e9e392f3b4-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.861607 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-config-data" (OuterVolumeSpecName: "config-data") pod "e536dcbd-a48d-409a-8154-49e9e392f3b4" (UID: "e536dcbd-a48d-409a-8154-49e9e392f3b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:00:07 crc kubenswrapper[4796]: I0930 17:00:07.933209 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e536dcbd-a48d-409a-8154-49e9e392f3b4-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.220599 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59"] Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.230849 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320815-bsb59"] Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.360571 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.360563 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e536dcbd-a48d-409a-8154-49e9e392f3b4","Type":"ContainerDied","Data":"11d8abd3345cf2814d4027c78bffa5e62cdc0e21792b584e26de4039079c00dc"} Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.361055 4796 scope.go:117] "RemoveContainer" containerID="043fbd9e40311e9722ca91e114c53df201a8be05d283a7698aa2de41c613b159" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.376877 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"65d8e2ee-85e4-4ea5-82ad-5de51cb4e427","Type":"ContainerStarted","Data":"d6b92442d6303a0246a13d37c41748bf0c24280899a38c42f2929841859d36b0"} Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.404640 4796 scope.go:117] "RemoveContainer" containerID="42a6b5cb9d6b4b85fe799eedf94866ca97de90de1b559246f517c4346e7e041d" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.421423 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.5267581850000003 podStartE2EDuration="42.421387095s" podCreationTimestamp="2025-09-30 16:59:26 +0000 UTC" firstStartedPulling="2025-09-30 16:59:28.228519575 +0000 UTC m=+2860.241798102" lastFinishedPulling="2025-09-30 17:00:07.123148485 +0000 UTC m=+2899.136427012" observedRunningTime="2025-09-30 17:00:08.417045339 +0000 UTC m=+2900.430323866" watchObservedRunningTime="2025-09-30 17:00:08.421387095 +0000 UTC m=+2900.434665642" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.444705 4796 scope.go:117] "RemoveContainer" containerID="805681bc59c000c3d10eafcf08305f67ed12c5d256a6d8ffbb4eb0a1a6d7a2ab" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.448319 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.465177 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.476378 4796 scope.go:117] "RemoveContainer" containerID="59877affe9f0dcd0279809ecef684a1b9996fd17cc7da6da6dd7ec8ec67e2e88" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.481770 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 17:00:08 crc kubenswrapper[4796]: E0930 17:00:08.482310 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-central-agent" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482334 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-central-agent" Sep 30 17:00:08 crc kubenswrapper[4796]: E0930 17:00:08.482349 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4052d15-4933-47cb-b44c-eaf792717a77" containerName="collect-profiles" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482358 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4052d15-4933-47cb-b44c-eaf792717a77" containerName="collect-profiles" Sep 30 17:00:08 crc kubenswrapper[4796]: E0930 17:00:08.482381 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-notification-agent" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482389 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-notification-agent" Sep 30 17:00:08 crc kubenswrapper[4796]: E0930 17:00:08.482411 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="sg-core" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482419 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="sg-core" Sep 30 17:00:08 crc kubenswrapper[4796]: E0930 17:00:08.482446 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="proxy-httpd" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482454 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="proxy-httpd" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482681 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-notification-agent" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482699 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="proxy-httpd" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482713 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="sg-core" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482732 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" containerName="ceilometer-central-agent" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.482754 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4052d15-4933-47cb-b44c-eaf792717a77" containerName="collect-profiles" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.485200 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.488646 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-28x5n" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.490611 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.491155 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.493140 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.545689 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cg4l\" (UniqueName: \"kubernetes.io/projected/97546445-acf2-4b62-936d-3c2d7fdeca87-kube-api-access-2cg4l\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.545818 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-scripts\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.545862 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-config-data\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.545899 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97546445-acf2-4b62-936d-3c2d7fdeca87-run-httpd\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.545922 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.545968 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97546445-acf2-4b62-936d-3c2d7fdeca87-log-httpd\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.647822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-scripts\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.648133 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-config-data\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.648211 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97546445-acf2-4b62-936d-3c2d7fdeca87-run-httpd\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.648244 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.648332 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97546445-acf2-4b62-936d-3c2d7fdeca87-log-httpd\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.648426 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cg4l\" (UniqueName: \"kubernetes.io/projected/97546445-acf2-4b62-936d-3c2d7fdeca87-kube-api-access-2cg4l\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.650651 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97546445-acf2-4b62-936d-3c2d7fdeca87-run-httpd\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.650657 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97546445-acf2-4b62-936d-3c2d7fdeca87-log-httpd\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.655334 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-config-data\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.655678 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.656613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97546445-acf2-4b62-936d-3c2d7fdeca87-scripts\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.672460 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:08 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:08 crc kubenswrapper[4796]: > Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.681973 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cg4l\" (UniqueName: \"kubernetes.io/projected/97546445-acf2-4b62-936d-3c2d7fdeca87-kube-api-access-2cg4l\") pod \"ceilometer-0\" (UID: \"97546445-acf2-4b62-936d-3c2d7fdeca87\") " pod="openstack/ceilometer-0" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.741926 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:00:08 crc kubenswrapper[4796]: E0930 17:00:08.742853 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.758568 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e43d3f-51c9-4caf-a2b4-394b0f4729f5" path="/var/lib/kubelet/pods/b0e43d3f-51c9-4caf-a2b4-394b0f4729f5/volumes" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.759793 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e536dcbd-a48d-409a-8154-49e9e392f3b4" path="/var/lib/kubelet/pods/e536dcbd-a48d-409a-8154-49e9e392f3b4/volumes" Sep 30 17:00:08 crc kubenswrapper[4796]: I0930 17:00:08.819863 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 17:00:09 crc kubenswrapper[4796]: I0930 17:00:09.310423 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 17:00:09 crc kubenswrapper[4796]: I0930 17:00:09.391955 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97546445-acf2-4b62-936d-3c2d7fdeca87","Type":"ContainerStarted","Data":"0dd9a7dbd6c95365769088972252b527f888da30094c756fd2db33dd2c5add9c"} Sep 30 17:00:10 crc kubenswrapper[4796]: I0930 17:00:10.404704 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97546445-acf2-4b62-936d-3c2d7fdeca87","Type":"ContainerStarted","Data":"7bfdead4c7464623a9af1afb82a8be65b07781ecbbc32ac66f6359f6a3eac300"} Sep 30 17:00:10 crc kubenswrapper[4796]: I0930 17:00:10.797226 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 17:00:10 crc kubenswrapper[4796]: I0930 17:00:10.798180 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:00:10 crc kubenswrapper[4796]: E0930 17:00:10.798494 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.375318 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.376393 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:00:11 crc kubenswrapper[4796]: E0930 17:00:11.376685 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.416352 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97546445-acf2-4b62-936d-3c2d7fdeca87","Type":"ContainerStarted","Data":"49188c749bdaeec6478e0b18965854442af1f0fa46584ec3ae9774b685b9ead7"} Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.680062 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:11 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:11 crc kubenswrapper[4796]: > Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.680150 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.680969 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"bfeff6e6601efe13b5d72cab54d22002eb465d32c1d0d581063bb3af5d5bf6ca"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:00:11 crc kubenswrapper[4796]: I0930 17:00:11.681030 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://bfeff6e6601efe13b5d72cab54d22002eb465d32c1d0d581063bb3af5d5bf6ca" gracePeriod=30 Sep 30 17:00:12 crc kubenswrapper[4796]: I0930 17:00:12.433849 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="bfeff6e6601efe13b5d72cab54d22002eb465d32c1d0d581063bb3af5d5bf6ca" exitCode=0 Sep 30 17:00:12 crc kubenswrapper[4796]: I0930 17:00:12.433908 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"bfeff6e6601efe13b5d72cab54d22002eb465d32c1d0d581063bb3af5d5bf6ca"} Sep 30 17:00:12 crc kubenswrapper[4796]: I0930 17:00:12.449971 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97546445-acf2-4b62-936d-3c2d7fdeca87","Type":"ContainerStarted","Data":"03c45318a36c066abfe2dc477b409b15c236c180c3091d983831ed3bbedb3178"} Sep 30 17:00:12 crc kubenswrapper[4796]: I0930 17:00:12.620782 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Sep 30 17:00:12 crc kubenswrapper[4796]: I0930 17:00:12.621670 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Sep 30 17:00:12 crc kubenswrapper[4796]: I0930 17:00:12.623670 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Sep 30 17:00:13 crc kubenswrapper[4796]: I0930 17:00:13.467781 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"6102f524aaa4eb191f145b032197be5383dde7b3a30691616e558783e9ae0f95"} Sep 30 17:00:13 crc kubenswrapper[4796]: I0930 17:00:13.474160 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97546445-acf2-4b62-936d-3c2d7fdeca87","Type":"ContainerStarted","Data":"ab488165564a7ce744975dc0b57988e096b6dd4c795d57e2470969aa9d8afc60"} Sep 30 17:00:13 crc kubenswrapper[4796]: I0930 17:00:13.474287 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 17:00:13 crc kubenswrapper[4796]: I0930 17:00:13.477072 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Sep 30 17:00:13 crc kubenswrapper[4796]: I0930 17:00:13.519954 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.184322625 podStartE2EDuration="5.519928729s" podCreationTimestamp="2025-09-30 17:00:08 +0000 UTC" firstStartedPulling="2025-09-30 17:00:09.310221779 +0000 UTC m=+2901.323500306" lastFinishedPulling="2025-09-30 17:00:12.645827883 +0000 UTC m=+2904.659106410" observedRunningTime="2025-09-30 17:00:13.511252568 +0000 UTC m=+2905.524531085" watchObservedRunningTime="2025-09-30 17:00:13.519928729 +0000 UTC m=+2905.533207266" Sep 30 17:00:17 crc kubenswrapper[4796]: I0930 17:00:17.635706 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:17 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:17 crc kubenswrapper[4796]: > Sep 30 17:00:19 crc kubenswrapper[4796]: I0930 17:00:19.734253 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:00:19 crc kubenswrapper[4796]: E0930 17:00:19.735338 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:00:20 crc kubenswrapper[4796]: I0930 17:00:20.716341 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:20 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:20 crc kubenswrapper[4796]: > Sep 30 17:00:23 crc kubenswrapper[4796]: I0930 17:00:23.763891 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:23 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:23 crc kubenswrapper[4796]: > Sep 30 17:00:23 crc kubenswrapper[4796]: I0930 17:00:23.764419 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:00:23 crc kubenswrapper[4796]: I0930 17:00:23.765331 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"6102f524aaa4eb191f145b032197be5383dde7b3a30691616e558783e9ae0f95"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:00:23 crc kubenswrapper[4796]: I0930 17:00:23.765372 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://6102f524aaa4eb191f145b032197be5383dde7b3a30691616e558783e9ae0f95" gracePeriod=30 Sep 30 17:00:24 crc kubenswrapper[4796]: I0930 17:00:24.604833 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="6102f524aaa4eb191f145b032197be5383dde7b3a30691616e558783e9ae0f95" exitCode=0 Sep 30 17:00:24 crc kubenswrapper[4796]: I0930 17:00:24.604902 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"6102f524aaa4eb191f145b032197be5383dde7b3a30691616e558783e9ae0f95"} Sep 30 17:00:24 crc kubenswrapper[4796]: I0930 17:00:24.605400 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"b31150de3714179bad68c9fbc9dd25aeb102ffdc9e2d186bf9cd69a5585c65bb"} Sep 30 17:00:24 crc kubenswrapper[4796]: I0930 17:00:24.605424 4796 scope.go:117] "RemoveContainer" containerID="bfeff6e6601efe13b5d72cab54d22002eb465d32c1d0d581063bb3af5d5bf6ca" Sep 30 17:00:25 crc kubenswrapper[4796]: I0930 17:00:25.733675 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:00:25 crc kubenswrapper[4796]: I0930 17:00:25.735198 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:00:25 crc kubenswrapper[4796]: E0930 17:00:25.735396 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:00:25 crc kubenswrapper[4796]: E0930 17:00:25.735436 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:00:29 crc kubenswrapper[4796]: I0930 17:00:29.740278 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:29 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:29 crc kubenswrapper[4796]: > Sep 30 17:00:32 crc kubenswrapper[4796]: I0930 17:00:32.651138 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:32 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:32 crc kubenswrapper[4796]: > Sep 30 17:00:32 crc kubenswrapper[4796]: I0930 17:00:32.733001 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:00:32 crc kubenswrapper[4796]: E0930 17:00:32.733541 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:00:33 crc kubenswrapper[4796]: I0930 17:00:33.206722 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Sep 30 17:00:35 crc kubenswrapper[4796]: I0930 17:00:35.609325 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:35 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:35 crc kubenswrapper[4796]: > Sep 30 17:00:35 crc kubenswrapper[4796]: I0930 17:00:35.609763 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:00:35 crc kubenswrapper[4796]: I0930 17:00:35.610480 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"b31150de3714179bad68c9fbc9dd25aeb102ffdc9e2d186bf9cd69a5585c65bb"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:00:35 crc kubenswrapper[4796]: I0930 17:00:35.610516 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://b31150de3714179bad68c9fbc9dd25aeb102ffdc9e2d186bf9cd69a5585c65bb" gracePeriod=30 Sep 30 17:00:36 crc kubenswrapper[4796]: I0930 17:00:36.726871 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="b31150de3714179bad68c9fbc9dd25aeb102ffdc9e2d186bf9cd69a5585c65bb" exitCode=0 Sep 30 17:00:36 crc kubenswrapper[4796]: I0930 17:00:36.726948 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"b31150de3714179bad68c9fbc9dd25aeb102ffdc9e2d186bf9cd69a5585c65bb"} Sep 30 17:00:36 crc kubenswrapper[4796]: I0930 17:00:36.727309 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6"} Sep 30 17:00:36 crc kubenswrapper[4796]: I0930 17:00:36.727345 4796 scope.go:117] "RemoveContainer" containerID="6102f524aaa4eb191f145b032197be5383dde7b3a30691616e558783e9ae0f95" Sep 30 17:00:36 crc kubenswrapper[4796]: I0930 17:00:36.733608 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:00:36 crc kubenswrapper[4796]: E0930 17:00:36.733932 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:00:38 crc kubenswrapper[4796]: I0930 17:00:38.740312 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:00:38 crc kubenswrapper[4796]: E0930 17:00:38.741180 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:00:38 crc kubenswrapper[4796]: I0930 17:00:38.830405 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 17:00:41 crc kubenswrapper[4796]: I0930 17:00:41.613605 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:41 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:41 crc kubenswrapper[4796]: > Sep 30 17:00:44 crc kubenswrapper[4796]: I0930 17:00:44.626290 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:44 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:44 crc kubenswrapper[4796]: > Sep 30 17:00:44 crc kubenswrapper[4796]: I0930 17:00:44.733659 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:00:44 crc kubenswrapper[4796]: E0930 17:00:44.734266 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:00:46 crc kubenswrapper[4796]: I0930 17:00:46.907749 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_eefa8387-6e63-41c6-b4c1-c3d6a994dfde/init-config-reloader/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.075553 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_eefa8387-6e63-41c6-b4c1-c3d6a994dfde/init-config-reloader/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.084323 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_eefa8387-6e63-41c6-b4c1-c3d6a994dfde/alertmanager/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.142507 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_eefa8387-6e63-41c6-b4c1-c3d6a994dfde/config-reloader/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.300224 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57cffd5bb6-xzdml_c7ed35b2-c6fc-4ab6-8054-d0a58f403425/barbican-api/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.392374 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57cffd5bb6-xzdml_c7ed35b2-c6fc-4ab6-8054-d0a58f403425/barbican-api-log/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.637993 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d5c775bd6-kffv8_919b1e9c-ab87-4902-a8e7-cad17ec22883/barbican-keystone-listener-log/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.660219 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d5c775bd6-kffv8_919b1e9c-ab87-4902-a8e7-cad17ec22883/barbican-keystone-listener/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.919013 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8568546bbc-6fvmg_d008ca00-46d2-4985-8f9d-3d48774e9a85/barbican-worker/0.log" Sep 30 17:00:47 crc kubenswrapper[4796]: I0930 17:00:47.937991 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8568546bbc-6fvmg_d008ca00-46d2-4985-8f9d-3d48774e9a85/barbican-worker-log/0.log" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.246248 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:00:48 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:00:48 crc kubenswrapper[4796]: > Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.246325 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.246969 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.247050 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6" gracePeriod=30 Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.267346 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-vr8qj_c62ce10b-8219-4394-b6c4-4514b202da42/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.470555 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_97546445-acf2-4b62-936d-3c2d7fdeca87/ceilometer-central-agent/0.log" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.520828 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_97546445-acf2-4b62-936d-3c2d7fdeca87/proxy-httpd/0.log" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.535091 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_97546445-acf2-4b62-936d-3c2d7fdeca87/ceilometer-notification-agent/0.log" Sep 30 17:00:48 crc kubenswrapper[4796]: I0930 17:00:48.932371 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_97546445-acf2-4b62-936d-3c2d7fdeca87/sg-core/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.078931 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_89e2552e-914c-48e2-99c1-06b6ef87aefe/cinder-api/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.134893 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_89e2552e-914c-48e2-99c1-06b6ef87aefe/cinder-api-log/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: E0930 17:00:49.276772 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.333095 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cf5d97fd-5337-41e8-9480-86b0342e9598/cinder-scheduler/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.353135 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cf5d97fd-5337-41e8-9480-86b0342e9598/probe/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.563379 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_dd002cf0-64b3-4fae-8624-a7114c4ea6bf/cloudkitty-api/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.617223 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_dd002cf0-64b3-4fae-8624-a7114c4ea6bf/cloudkitty-api-log/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.733579 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:00:49 crc kubenswrapper[4796]: E0930 17:00:49.733859 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.797962 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-db-create-rpkc5_7eb4171f-878b-4a0e-8852-99700fedd6ea/mariadb-database-create/0.log" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.866304 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6" exitCode=0 Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.866357 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6"} Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.866399 4796 scope.go:117] "RemoveContainer" containerID="b31150de3714179bad68c9fbc9dd25aeb102ffdc9e2d186bf9cd69a5585c65bb" Sep 30 17:00:49 crc kubenswrapper[4796]: I0930 17:00:49.867199 4796 scope.go:117] "RemoveContainer" containerID="2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6" Sep 30 17:00:49 crc kubenswrapper[4796]: E0930 17:00:49.867491 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.031480 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-db-sync-glbqt_8fd89bba-353c-4e91-8821-1ed2d35e2507/cloudkitty-db-sync/0.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.057131 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-fa82-account-create-q9fh4_a2d3ff67-b482-4050-baad-14165c0f626a/mariadb-account-create/0.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.269635 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-bccccd5f6-j4nk9_e8e35857-aed6-49cf-8ede-715b5b915ffe/loki-distributor/0.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.275298 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_0e12debc-3efb-4e67-9e07-2531628e41ab/loki-compactor/0.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.521313 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-89dc74b89-cpv5x_165eb10b-4c60-4635-928e-c04a72e9b4f5/gateway/4.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.521500 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-89dc74b89-cpv5x_165eb10b-4c60-4635-928e-c04a72e9b4f5/gateway/4.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.674163 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-89dc74b89-hv6jn_15e69566-5b25-4929-a40b-1a3e2d710800/gateway/4.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.733694 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-89dc74b89-hv6jn_15e69566-5b25-4929-a40b-1a3e2d710800/gateway/4.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.833921 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_048ee9b2-bfc7-4f0c-ae5b-1fab51d7dc0c/loki-index-gateway/0.log" Sep 30 17:00:50 crc kubenswrapper[4796]: I0930 17:00:50.973306 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_6cbd3fc7-5956-4320-b074-bc791068c9cc/loki-ingester/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.068699 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-6b6cdc96db-2rjz6_914424ae-a0ec-4578-9a57-824fceadcf0b/loki-querier/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.179119 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-7c7846d94c-6hm8f_b18114f1-3f03-4f73-b46e-62056032d351/loki-query-frontend/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.295826 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad/cloudkitty-proc/3.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.387290 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad/cloudkitty-proc/3.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.482724 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-storageinit-5s8tn_d96b76a9-dd08-43c3-a7e8-2bc542706345/cloudkitty-storageinit/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.598634 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-h4sxw_6ba38983-af8e-47bc-9081-7404ef0dfefc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.764937 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-pgvb7_27c8d7f6-0fb7-48cb-9668-d2aabf4c34c2/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.840479 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-r94sx_2c6044b7-9944-42ac-821a-8ddad26ad823/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:51 crc kubenswrapper[4796]: I0930 17:00:51.982879 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-f4rxb_eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d/init/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.516730 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-f4rxb_eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d/init/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.533165 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-bmmqc_2d41515e-6a93-42d8-887c-c4f0f748bf3f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.538548 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-f4rxb_eb4340a6-f8f3-4d95-922c-fddf4e6c5a9d/dnsmasq-dns/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.704211 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4cf6a891-18be-4f28-ad83-8a1f399d1cbd/glance-httpd/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.720407 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4cf6a891-18be-4f28-ad83-8a1f399d1cbd/glance-log/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.888378 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_63bfb928-2342-4432-be94-4f37611ae51e/glance-httpd/0.log" Sep 30 17:00:52 crc kubenswrapper[4796]: I0930 17:00:52.934271 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_63bfb928-2342-4432-be94-4f37611ae51e/glance-log/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.075948 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lbvlc_185ba20f-66a9-4425-a1ba-bfc42a8f1c01/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.172972 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zm5lv_743c014a-a242-4a7a-96e5-167a0d0928fb/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.408113 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75d78779f8-w7h2x_624ee0f3-fc1d-4e41-aaae-74767b105143/keystone-api/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.452658 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-89pdx_0f40019c-f145-46f4-89f3-cc7244d69faf/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.689069 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-685c9578b5-5pzck_3d40c196-945d-4f46-90ca-ea6b287a0f7c/neutron-api/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.733882 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:00:53 crc kubenswrapper[4796]: E0930 17:00:53.734151 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.825332 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-685c9578b5-5pzck_3d40c196-945d-4f46-90ca-ea6b287a0f7c/neutron-httpd/0.log" Sep 30 17:00:53 crc kubenswrapper[4796]: I0930 17:00:53.890374 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-5n7ds_f90058d7-bebf-4a04-8fb5-045010f7a6b3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:54 crc kubenswrapper[4796]: I0930 17:00:54.319340 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_dec95fef-d221-4739-849e-5e071ce2311a/nova-api-log/0.log" Sep 30 17:00:54 crc kubenswrapper[4796]: I0930 17:00:54.383750 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_dec95fef-d221-4739-849e-5e071ce2311a/nova-api-api/0.log" Sep 30 17:00:54 crc kubenswrapper[4796]: I0930 17:00:54.722352 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_875eff01-1202-4914-8f8b-172f661da26a/nova-cell0-conductor-conductor/0.log" Sep 30 17:00:54 crc kubenswrapper[4796]: I0930 17:00:54.751710 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_d5b132cf-1062-45b8-a1b2-8626de3f640d/nova-cell1-conductor-conductor/0.log" Sep 30 17:00:55 crc kubenswrapper[4796]: I0930 17:00:55.000856 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8a76e9ab-55ec-4394-94b3-668b7b146c81/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 17:00:55 crc kubenswrapper[4796]: I0930 17:00:55.129567 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-6gb98_6212b5a6-1cb0-44ec-bd7c-46ae527d9fe3/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:55 crc kubenswrapper[4796]: I0930 17:00:55.351837 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a679bcdb-21e4-4761-adc7-1c90e79c6650/nova-metadata-log/0.log" Sep 30 17:00:55 crc kubenswrapper[4796]: I0930 17:00:55.734077 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a6185e29-465b-4ab1-b963-88ff87922e39/nova-scheduler-scheduler/0.log" Sep 30 17:00:56 crc kubenswrapper[4796]: I0930 17:00:56.253046 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2bb8f271-8840-42e7-984c-31cf8fc8bea0/mysql-bootstrap/0.log" Sep 30 17:00:56 crc kubenswrapper[4796]: I0930 17:00:56.370034 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2bb8f271-8840-42e7-984c-31cf8fc8bea0/mysql-bootstrap/0.log" Sep 30 17:00:56 crc kubenswrapper[4796]: I0930 17:00:56.426510 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a679bcdb-21e4-4761-adc7-1c90e79c6650/nova-metadata-metadata/0.log" Sep 30 17:00:56 crc kubenswrapper[4796]: I0930 17:00:56.504174 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2bb8f271-8840-42e7-984c-31cf8fc8bea0/galera/0.log" Sep 30 17:00:56 crc kubenswrapper[4796]: I0930 17:00:56.673876 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6162abb2-dddc-4cb6-bd50-0e71ae9534bd/mysql-bootstrap/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.007355 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6162abb2-dddc-4cb6-bd50-0e71ae9534bd/mysql-bootstrap/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.022834 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6162abb2-dddc-4cb6-bd50-0e71ae9534bd/galera/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.180447 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1df486ab-b2b8-471d-b88c-c800f254716c/openstackclient/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.338117 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-pmjrt_f5bcbf15-75a7-4b1f-bb3d-4d58f59de746/openstack-network-exporter/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.576546 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lg6m7_403ec30e-d599-4940-8eaf-6cccb3200ed9/ovsdb-server-init/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.809817 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lg6m7_403ec30e-d599-4940-8eaf-6cccb3200ed9/ovsdb-server/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.821225 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lg6m7_403ec30e-d599-4940-8eaf-6cccb3200ed9/ovsdb-server-init/0.log" Sep 30 17:00:57 crc kubenswrapper[4796]: I0930 17:00:57.872541 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lg6m7_403ec30e-d599-4940-8eaf-6cccb3200ed9/ovs-vswitchd/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.066570 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vfdx6_713f98aa-c3ec-4627-bc90-a1a577ca1c27/ovn-controller/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.356590 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vtt88_abb53121-c532-478a-9670-8350dd3e477b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.378301 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_43d2a8c3-5a3c-475d-a348-843d5f59b008/openstack-network-exporter/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.629862 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_43d2a8c3-5a3c-475d-a348-843d5f59b008/ovn-northd/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.649317 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_85b72cb4-9a34-4d07-8b11-dc60f4d540c9/openstack-network-exporter/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.912239 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_85b72cb4-9a34-4d07-8b11-dc60f4d540c9/ovsdbserver-nb/0.log" Sep 30 17:00:58 crc kubenswrapper[4796]: I0930 17:00:58.938186 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a121b29f-66c2-4b9f-99b2-9a6c4a46473f/openstack-network-exporter/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.110380 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a121b29f-66c2-4b9f-99b2-9a6c4a46473f/ovsdbserver-sb/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.204210 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5785c489dd-mzgjb_2a8a5505-564d-4173-8f22-869b0bec6e2c/placement-api/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.402275 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5785c489dd-mzgjb_2a8a5505-564d-4173-8f22-869b0bec6e2c/placement-log/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.492303 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_65d8e2ee-85e4-4ea5-82ad-5de51cb4e427/init-config-reloader/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.720139 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_65d8e2ee-85e4-4ea5-82ad-5de51cb4e427/init-config-reloader/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.734659 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:00:59 crc kubenswrapper[4796]: E0930 17:00:59.734926 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.735428 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_65d8e2ee-85e4-4ea5-82ad-5de51cb4e427/config-reloader/0.log" Sep 30 17:00:59 crc kubenswrapper[4796]: I0930 17:00:59.761082 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_65d8e2ee-85e4-4ea5-82ad-5de51cb4e427/prometheus/0.log" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.235763 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29320861-7w25p"] Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.237881 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.283817 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320861-7w25p"] Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.334820 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_65d8e2ee-85e4-4ea5-82ad-5de51cb4e427/thanos-sidecar/0.log" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.341412 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d4dc01f-d8ed-4040-b4f7-53498e2cc237/setup-container/0.log" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.383198 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-combined-ca-bundle\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.383250 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk5zh\" (UniqueName: \"kubernetes.io/projected/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-kube-api-access-kk5zh\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.383453 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-config-data\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.383519 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-fernet-keys\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.484739 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk5zh\" (UniqueName: \"kubernetes.io/projected/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-kube-api-access-kk5zh\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.484907 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-config-data\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.484955 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-fernet-keys\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.485006 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-combined-ca-bundle\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.512868 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-combined-ca-bundle\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.513447 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-fernet-keys\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.513797 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-config-data\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.517392 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk5zh\" (UniqueName: \"kubernetes.io/projected/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-kube-api-access-kk5zh\") pod \"keystone-cron-29320861-7w25p\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.580877 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.606326 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d4dc01f-d8ed-4040-b4f7-53498e2cc237/setup-container/0.log" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.679804 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d4dc01f-d8ed-4040-b4f7-53498e2cc237/rabbitmq/0.log" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.747041 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:01:00 crc kubenswrapper[4796]: E0930 17:01:00.748997 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:01:00 crc kubenswrapper[4796]: I0930 17:01:00.913203 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5/setup-container/0.log" Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.155402 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320861-7w25p"] Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.173281 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5/setup-container/0.log" Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.194128 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7c7ff799-8fd6-4dd9-a2f4-9e0ab443d7a5/rabbitmq/0.log" Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.398309 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qnps2_8b67950f-98c0-44fe-a09a-77e356e069ff/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.433859 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-47dk6_80a695f7-4e2a-47f4-a4f4-724330d4e2e8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.748567 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6lrs8_6146b235-0d04-4d50-b28b-8531fe47a5d2/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:01:01 crc kubenswrapper[4796]: I0930 17:01:01.869240 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-q2j8x_1351ba7c-3e3e-42ff-960f-8ada3d7e490a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.034775 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320861-7w25p" event={"ID":"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4","Type":"ContainerStarted","Data":"2d1f6c939187102799ffc5f524e4f9c6dccdc9015bf2d978559eee52cb6237ae"} Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.035120 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320861-7w25p" event={"ID":"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4","Type":"ContainerStarted","Data":"51081421b191d6dffc9abf73caa17edb89e298650b0d02456d9269f718ca4edb"} Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.066809 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29320861-7w25p" podStartSLOduration=2.066786784 podStartE2EDuration="2.066786784s" podCreationTimestamp="2025-09-30 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 17:01:02.048400101 +0000 UTC m=+2954.061678628" watchObservedRunningTime="2025-09-30 17:01:02.066786784 +0000 UTC m=+2954.080065311" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.079517 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-j5xtv_35c5d020-3189-41c5-90f4-4544c251539e/ssh-known-hosts-edpm-deployment/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.322007 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85f459d7c9-l56gz_df49731d-3260-4e0a-8633-e2a7ca68011d/proxy-httpd/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.344688 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85f459d7c9-l56gz_df49731d-3260-4e0a-8633-e2a7ca68011d/proxy-server/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.667026 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m6xqp_420fc105-e180-4b89-af90-af130839e938/swift-ring-rebalance/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.753623 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/account-auditor/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.910172 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/account-replicator/0.log" Sep 30 17:01:02 crc kubenswrapper[4796]: I0930 17:01:02.922456 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/account-reaper/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.032912 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/account-server/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.126156 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/container-auditor/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.162194 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/container-replicator/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.296244 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/container-server/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.381520 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/container-updater/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.414008 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/object-auditor/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.499498 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/object-expirer/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.615178 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/object-replicator/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.632219 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/object-server/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.728430 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/object-updater/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.868926 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/swift-recon-cron/0.log" Sep 30 17:01:03 crc kubenswrapper[4796]: I0930 17:01:03.901590 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_24528035-6f6d-4269-9e88-6171795db8a7/rsync/0.log" Sep 30 17:01:04 crc kubenswrapper[4796]: I0930 17:01:04.136768 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-pmh2p_e8acd478-67ee-465a-b801-6731a62328f3/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:01:04 crc kubenswrapper[4796]: I0930 17:01:04.268804 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-b77pv_2e9c9f25-af36-4281-b747-d03a289f22f5/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 17:01:04 crc kubenswrapper[4796]: I0930 17:01:04.550134 4796 scope.go:117] "RemoveContainer" containerID="04ce5f567fa6307b9536918267c255b9b7cfca95a8e0ef1e694e35f8607eab4f" Sep 30 17:01:04 crc kubenswrapper[4796]: I0930 17:01:04.733698 4796 scope.go:117] "RemoveContainer" containerID="2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6" Sep 30 17:01:04 crc kubenswrapper[4796]: E0930 17:01:04.734241 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:01:06 crc kubenswrapper[4796]: I0930 17:01:06.079714 4796 generic.go:334] "Generic (PLEG): container finished" podID="cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" containerID="2d1f6c939187102799ffc5f524e4f9c6dccdc9015bf2d978559eee52cb6237ae" exitCode=0 Sep 30 17:01:06 crc kubenswrapper[4796]: I0930 17:01:06.079781 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320861-7w25p" event={"ID":"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4","Type":"ContainerDied","Data":"2d1f6c939187102799ffc5f524e4f9c6dccdc9015bf2d978559eee52cb6237ae"} Sep 30 17:01:06 crc kubenswrapper[4796]: I0930 17:01:06.733813 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:01:06 crc kubenswrapper[4796]: E0930 17:01:06.734068 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.157400 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_94a4df46-32b9-4a28-848e-4c90a2660b82/memcached/0.log" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.532005 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.636606 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-fernet-keys\") pod \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.636923 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-combined-ca-bundle\") pod \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.637010 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-config-data\") pod \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.637110 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk5zh\" (UniqueName: \"kubernetes.io/projected/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-kube-api-access-kk5zh\") pod \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\" (UID: \"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4\") " Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.643659 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-kube-api-access-kk5zh" (OuterVolumeSpecName: "kube-api-access-kk5zh") pod "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" (UID: "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4"). InnerVolumeSpecName "kube-api-access-kk5zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.650098 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" (UID: "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.693474 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" (UID: "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.731876 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-config-data" (OuterVolumeSpecName: "config-data") pod "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" (UID: "cd4a0d1e-3090-4d7f-89a3-780a8c2670c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.740713 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk5zh\" (UniqueName: \"kubernetes.io/projected/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-kube-api-access-kk5zh\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.740748 4796 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.740757 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:07 crc kubenswrapper[4796]: I0930 17:01:07.740765 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a0d1e-3090-4d7f-89a3-780a8c2670c4-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:08 crc kubenswrapper[4796]: I0930 17:01:08.105449 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320861-7w25p" event={"ID":"cd4a0d1e-3090-4d7f-89a3-780a8c2670c4","Type":"ContainerDied","Data":"51081421b191d6dffc9abf73caa17edb89e298650b0d02456d9269f718ca4edb"} Sep 30 17:01:08 crc kubenswrapper[4796]: I0930 17:01:08.105500 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51081421b191d6dffc9abf73caa17edb89e298650b0d02456d9269f718ca4edb" Sep 30 17:01:08 crc kubenswrapper[4796]: I0930 17:01:08.105560 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320861-7w25p" Sep 30 17:01:13 crc kubenswrapper[4796]: I0930 17:01:13.733397 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:01:13 crc kubenswrapper[4796]: I0930 17:01:13.734211 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:01:13 crc kubenswrapper[4796]: E0930 17:01:13.734293 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:01:13 crc kubenswrapper[4796]: E0930 17:01:13.734391 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:01:18 crc kubenswrapper[4796]: I0930 17:01:18.742366 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:01:18 crc kubenswrapper[4796]: I0930 17:01:18.742862 4796 scope.go:117] "RemoveContainer" containerID="2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6" Sep 30 17:01:19 crc kubenswrapper[4796]: I0930 17:01:19.234716 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"25855246f7beddcb03c5746fc2f225e9ce5ec312915f23a02820182d348a4c4f"} Sep 30 17:01:19 crc kubenswrapper[4796]: I0930 17:01:19.236287 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerStarted","Data":"ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1"} Sep 30 17:01:19 crc kubenswrapper[4796]: I0930 17:01:19.236622 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 17:01:19 crc kubenswrapper[4796]: I0930 17:01:19.239117 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.254:8081/ready\": dial tcp 10.217.0.254:8081: connect: connection refused" Sep 30 17:01:19 crc kubenswrapper[4796]: I0930 17:01:19.279594 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podStartSLOduration=180.450916377 podStartE2EDuration="3m9.279576076s" podCreationTimestamp="2025-09-30 16:58:10 +0000 UTC" firstStartedPulling="2025-09-30 16:58:11.496147675 +0000 UTC m=+2783.509426202" lastFinishedPulling="2025-09-30 16:58:20.324807374 +0000 UTC m=+2792.338085901" observedRunningTime="2025-09-30 17:01:19.273024646 +0000 UTC m=+2971.286303183" watchObservedRunningTime="2025-09-30 17:01:19.279576076 +0000 UTC m=+2971.292854603" Sep 30 17:01:20 crc kubenswrapper[4796]: I0930 17:01:20.248010 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" exitCode=2 Sep 30 17:01:20 crc kubenswrapper[4796]: I0930 17:01:20.248048 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1"} Sep 30 17:01:20 crc kubenswrapper[4796]: I0930 17:01:20.248078 4796 scope.go:117] "RemoveContainer" containerID="518cbc44237f99baf1ff9c2548125d4b1dfdc6bdb7b4e9c8948a800751facd53" Sep 30 17:01:20 crc kubenswrapper[4796]: I0930 17:01:20.248803 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:01:20 crc kubenswrapper[4796]: E0930 17:01:20.249286 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:01:21 crc kubenswrapper[4796]: I0930 17:01:21.261130 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:01:21 crc kubenswrapper[4796]: E0930 17:01:21.261704 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:01:23 crc kubenswrapper[4796]: I0930 17:01:23.584764 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:01:23 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:01:23 crc kubenswrapper[4796]: > Sep 30 17:01:24 crc kubenswrapper[4796]: I0930 17:01:24.734377 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:01:24 crc kubenswrapper[4796]: E0930 17:01:24.734961 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:01:26 crc kubenswrapper[4796]: I0930 17:01:26.624383 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:01:26 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:01:26 crc kubenswrapper[4796]: > Sep 30 17:01:28 crc kubenswrapper[4796]: I0930 17:01:28.742462 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.352507 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" exitCode=2 Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.352606 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641"} Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.352837 4796 scope.go:117] "RemoveContainer" containerID="fa6c886d7acdad0c210143345105069f187342c3b018f8571181ca27b4d01b26" Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.353512 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:01:29 crc kubenswrapper[4796]: E0930 17:01:29.353745 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.633125 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:01:29 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:01:29 crc kubenswrapper[4796]: > Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.633211 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.633934 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"25855246f7beddcb03c5746fc2f225e9ce5ec312915f23a02820182d348a4c4f"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:01:29 crc kubenswrapper[4796]: I0930 17:01:29.633966 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://25855246f7beddcb03c5746fc2f225e9ce5ec312915f23a02820182d348a4c4f" gracePeriod=30 Sep 30 17:01:30 crc kubenswrapper[4796]: I0930 17:01:30.366254 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="25855246f7beddcb03c5746fc2f225e9ce5ec312915f23a02820182d348a4c4f" exitCode=0 Sep 30 17:01:30 crc kubenswrapper[4796]: I0930 17:01:30.366328 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"25855246f7beddcb03c5746fc2f225e9ce5ec312915f23a02820182d348a4c4f"} Sep 30 17:01:30 crc kubenswrapper[4796]: I0930 17:01:30.366624 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4"} Sep 30 17:01:30 crc kubenswrapper[4796]: I0930 17:01:30.366653 4796 scope.go:117] "RemoveContainer" containerID="2f38cf73692ecf869c9d72b8c1de115892017f572de911a094845a6c370d39c6" Sep 30 17:01:31 crc kubenswrapper[4796]: I0930 17:01:31.375387 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 17:01:31 crc kubenswrapper[4796]: I0930 17:01:31.378879 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:01:31 crc kubenswrapper[4796]: E0930 17:01:31.379454 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:01:33 crc kubenswrapper[4796]: I0930 17:01:33.733320 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:01:33 crc kubenswrapper[4796]: E0930 17:01:33.733603 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:01:35 crc kubenswrapper[4796]: I0930 17:01:35.605253 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:01:35 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:01:35 crc kubenswrapper[4796]: > Sep 30 17:01:38 crc kubenswrapper[4796]: I0930 17:01:38.471824 4796 generic.go:334] "Generic (PLEG): container finished" podID="599fe9b7-954d-408e-b7b8-aa941ac23480" containerID="e21a17e48af95794a5fb3b5abbd4975743723673ffb0e75c19b211ada2d08712" exitCode=0 Sep 30 17:01:38 crc kubenswrapper[4796]: I0930 17:01:38.471904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" event={"ID":"599fe9b7-954d-408e-b7b8-aa941ac23480","Type":"ContainerDied","Data":"e21a17e48af95794a5fb3b5abbd4975743723673ffb0e75c19b211ada2d08712"} Sep 30 17:01:38 crc kubenswrapper[4796]: I0930 17:01:38.623519 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:01:38 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:01:38 crc kubenswrapper[4796]: > Sep 30 17:01:38 crc kubenswrapper[4796]: I0930 17:01:38.745022 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:01:38 crc kubenswrapper[4796]: E0930 17:01:38.745409 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.633141 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.636621 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599fe9b7-954d-408e-b7b8-aa941ac23480-host\") pod \"599fe9b7-954d-408e-b7b8-aa941ac23480\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.636710 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/599fe9b7-954d-408e-b7b8-aa941ac23480-host" (OuterVolumeSpecName: "host") pod "599fe9b7-954d-408e-b7b8-aa941ac23480" (UID: "599fe9b7-954d-408e-b7b8-aa941ac23480"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.636888 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n658r\" (UniqueName: \"kubernetes.io/projected/599fe9b7-954d-408e-b7b8-aa941ac23480-kube-api-access-n658r\") pod \"599fe9b7-954d-408e-b7b8-aa941ac23480\" (UID: \"599fe9b7-954d-408e-b7b8-aa941ac23480\") " Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.637560 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599fe9b7-954d-408e-b7b8-aa941ac23480-host\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.647597 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/599fe9b7-954d-408e-b7b8-aa941ac23480-kube-api-access-n658r" (OuterVolumeSpecName: "kube-api-access-n658r") pod "599fe9b7-954d-408e-b7b8-aa941ac23480" (UID: "599fe9b7-954d-408e-b7b8-aa941ac23480"). InnerVolumeSpecName "kube-api-access-n658r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.672466 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-sc8t9"] Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.684432 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-sc8t9"] Sep 30 17:01:39 crc kubenswrapper[4796]: I0930 17:01:39.739322 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n658r\" (UniqueName: \"kubernetes.io/projected/599fe9b7-954d-408e-b7b8-aa941ac23480-kube-api-access-n658r\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.504713 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02f191c4aaf1f4c3a025bc83c33338ab693d717891ce57a36c798bbf4c12a1ee" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.504795 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-sc8t9" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.743836 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="599fe9b7-954d-408e-b7b8-aa941ac23480" path="/var/lib/kubelet/pods/599fe9b7-954d-408e-b7b8-aa941ac23480/volumes" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.796653 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.797870 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:01:40 crc kubenswrapper[4796]: E0930 17:01:40.798453 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.862491 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-7rcfs"] Sep 30 17:01:40 crc kubenswrapper[4796]: E0930 17:01:40.862895 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" containerName="keystone-cron" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.862908 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" containerName="keystone-cron" Sep 30 17:01:40 crc kubenswrapper[4796]: E0930 17:01:40.862920 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="599fe9b7-954d-408e-b7b8-aa941ac23480" containerName="container-00" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.862926 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="599fe9b7-954d-408e-b7b8-aa941ac23480" containerName="container-00" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.863158 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="599fe9b7-954d-408e-b7b8-aa941ac23480" containerName="container-00" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.863179 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4a0d1e-3090-4d7f-89a3-780a8c2670c4" containerName="keystone-cron" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.864520 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.867052 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rk8bx"/"default-dockercfg-zd74m" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.966141 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0b0141c8-9757-44ae-8b88-18b5d186abeb-host\") pod \"crc-debug-7rcfs\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:40 crc kubenswrapper[4796]: I0930 17:01:40.966330 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnrnh\" (UniqueName: \"kubernetes.io/projected/0b0141c8-9757-44ae-8b88-18b5d186abeb-kube-api-access-wnrnh\") pod \"crc-debug-7rcfs\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.068092 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnrnh\" (UniqueName: \"kubernetes.io/projected/0b0141c8-9757-44ae-8b88-18b5d186abeb-kube-api-access-wnrnh\") pod \"crc-debug-7rcfs\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.068475 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0b0141c8-9757-44ae-8b88-18b5d186abeb-host\") pod \"crc-debug-7rcfs\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.068659 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0b0141c8-9757-44ae-8b88-18b5d186abeb-host\") pod \"crc-debug-7rcfs\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.095228 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnrnh\" (UniqueName: \"kubernetes.io/projected/0b0141c8-9757-44ae-8b88-18b5d186abeb-kube-api-access-wnrnh\") pod \"crc-debug-7rcfs\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.187180 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:41 crc kubenswrapper[4796]: W0930 17:01:41.224345 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b0141c8_9757_44ae_8b88_18b5d186abeb.slice/crio-7874b96e099fbde2cde4506e3d81beffc7bf9f4f6e52fbc9543e26706ad42ac5 WatchSource:0}: Error finding container 7874b96e099fbde2cde4506e3d81beffc7bf9f4f6e52fbc9543e26706ad42ac5: Status 404 returned error can't find the container with id 7874b96e099fbde2cde4506e3d81beffc7bf9f4f6e52fbc9543e26706ad42ac5 Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.376035 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.377485 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:01:41 crc kubenswrapper[4796]: E0930 17:01:41.377913 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.520872 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" event={"ID":"0b0141c8-9757-44ae-8b88-18b5d186abeb","Type":"ContainerStarted","Data":"f8834f992b2ad1927b79ab593c9628bf18c6eddf805e59cfb79faa0c38eb55b5"} Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.520928 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" event={"ID":"0b0141c8-9757-44ae-8b88-18b5d186abeb","Type":"ContainerStarted","Data":"7874b96e099fbde2cde4506e3d81beffc7bf9f4f6e52fbc9543e26706ad42ac5"} Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.553878 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" podStartSLOduration=1.553854273 podStartE2EDuration="1.553854273s" podCreationTimestamp="2025-09-30 17:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 17:01:41.535510583 +0000 UTC m=+2993.548789110" watchObservedRunningTime="2025-09-30 17:01:41.553854273 +0000 UTC m=+2993.567132800" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.659561 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:01:41 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:01:41 crc kubenswrapper[4796]: > Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.659862 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.660725 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:01:41 crc kubenswrapper[4796]: I0930 17:01:41.660752 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" gracePeriod=30 Sep 30 17:01:42 crc kubenswrapper[4796]: E0930 17:01:42.082631 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:01:42 crc kubenswrapper[4796]: I0930 17:01:42.538665 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" exitCode=0 Sep 30 17:01:42 crc kubenswrapper[4796]: I0930 17:01:42.538714 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4"} Sep 30 17:01:42 crc kubenswrapper[4796]: I0930 17:01:42.538767 4796 scope.go:117] "RemoveContainer" containerID="25855246f7beddcb03c5746fc2f225e9ce5ec312915f23a02820182d348a4c4f" Sep 30 17:01:42 crc kubenswrapper[4796]: I0930 17:01:42.540059 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:01:42 crc kubenswrapper[4796]: E0930 17:01:42.540364 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:01:42 crc kubenswrapper[4796]: I0930 17:01:42.545226 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" event={"ID":"0b0141c8-9757-44ae-8b88-18b5d186abeb","Type":"ContainerDied","Data":"f8834f992b2ad1927b79ab593c9628bf18c6eddf805e59cfb79faa0c38eb55b5"} Sep 30 17:01:42 crc kubenswrapper[4796]: I0930 17:01:42.545207 4796 generic.go:334] "Generic (PLEG): container finished" podID="0b0141c8-9757-44ae-8b88-18b5d186abeb" containerID="f8834f992b2ad1927b79ab593c9628bf18c6eddf805e59cfb79faa0c38eb55b5" exitCode=0 Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.668682 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.816746 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0b0141c8-9757-44ae-8b88-18b5d186abeb-host\") pod \"0b0141c8-9757-44ae-8b88-18b5d186abeb\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.816801 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnrnh\" (UniqueName: \"kubernetes.io/projected/0b0141c8-9757-44ae-8b88-18b5d186abeb-kube-api-access-wnrnh\") pod \"0b0141c8-9757-44ae-8b88-18b5d186abeb\" (UID: \"0b0141c8-9757-44ae-8b88-18b5d186abeb\") " Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.816937 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b0141c8-9757-44ae-8b88-18b5d186abeb-host" (OuterVolumeSpecName: "host") pod "0b0141c8-9757-44ae-8b88-18b5d186abeb" (UID: "0b0141c8-9757-44ae-8b88-18b5d186abeb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.817833 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0b0141c8-9757-44ae-8b88-18b5d186abeb-host\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.826134 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b0141c8-9757-44ae-8b88-18b5d186abeb-kube-api-access-wnrnh" (OuterVolumeSpecName: "kube-api-access-wnrnh") pod "0b0141c8-9757-44ae-8b88-18b5d186abeb" (UID: "0b0141c8-9757-44ae-8b88-18b5d186abeb"). InnerVolumeSpecName "kube-api-access-wnrnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:01:43 crc kubenswrapper[4796]: I0930 17:01:43.919253 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnrnh\" (UniqueName: \"kubernetes.io/projected/0b0141c8-9757-44ae-8b88-18b5d186abeb-kube-api-access-wnrnh\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:44 crc kubenswrapper[4796]: I0930 17:01:44.571397 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" event={"ID":"0b0141c8-9757-44ae-8b88-18b5d186abeb","Type":"ContainerDied","Data":"7874b96e099fbde2cde4506e3d81beffc7bf9f4f6e52fbc9543e26706ad42ac5"} Sep 30 17:01:44 crc kubenswrapper[4796]: I0930 17:01:44.571441 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7874b96e099fbde2cde4506e3d81beffc7bf9f4f6e52fbc9543e26706ad42ac5" Sep 30 17:01:44 crc kubenswrapper[4796]: I0930 17:01:44.571461 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-7rcfs" Sep 30 17:01:48 crc kubenswrapper[4796]: I0930 17:01:48.052882 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-7rcfs"] Sep 30 17:01:48 crc kubenswrapper[4796]: I0930 17:01:48.061766 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-7rcfs"] Sep 30 17:01:48 crc kubenswrapper[4796]: I0930 17:01:48.768212 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b0141c8-9757-44ae-8b88-18b5d186abeb" path="/var/lib/kubelet/pods/0b0141c8-9757-44ae-8b88-18b5d186abeb/volumes" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.234106 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-8qqvf"] Sep 30 17:01:49 crc kubenswrapper[4796]: E0930 17:01:49.234677 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b0141c8-9757-44ae-8b88-18b5d186abeb" containerName="container-00" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.234694 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b0141c8-9757-44ae-8b88-18b5d186abeb" containerName="container-00" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.234971 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b0141c8-9757-44ae-8b88-18b5d186abeb" containerName="container-00" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.235899 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.239769 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rk8bx"/"default-dockercfg-zd74m" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.409525 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-host\") pod \"crc-debug-8qqvf\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.409641 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gnqz\" (UniqueName: \"kubernetes.io/projected/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-kube-api-access-5gnqz\") pod \"crc-debug-8qqvf\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.511121 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-host\") pod \"crc-debug-8qqvf\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.511252 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-host\") pod \"crc-debug-8qqvf\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.511267 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gnqz\" (UniqueName: \"kubernetes.io/projected/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-kube-api-access-5gnqz\") pod \"crc-debug-8qqvf\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.529705 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gnqz\" (UniqueName: \"kubernetes.io/projected/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-kube-api-access-5gnqz\") pod \"crc-debug-8qqvf\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.567835 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:49 crc kubenswrapper[4796]: W0930 17:01:49.598848 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5db2d5c0_ac0b_4d7a_bc88_ff7d53bc3e83.slice/crio-2bf922cc41d7ad910462cf38c603e53b334c8579dafc7fe664ae0282ae74fdb0 WatchSource:0}: Error finding container 2bf922cc41d7ad910462cf38c603e53b334c8579dafc7fe664ae0282ae74fdb0: Status 404 returned error can't find the container with id 2bf922cc41d7ad910462cf38c603e53b334c8579dafc7fe664ae0282ae74fdb0 Sep 30 17:01:49 crc kubenswrapper[4796]: I0930 17:01:49.631913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" event={"ID":"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83","Type":"ContainerStarted","Data":"2bf922cc41d7ad910462cf38c603e53b334c8579dafc7fe664ae0282ae74fdb0"} Sep 30 17:01:50 crc kubenswrapper[4796]: I0930 17:01:50.643626 4796 generic.go:334] "Generic (PLEG): container finished" podID="5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" containerID="32ba7d5aa8d1798bcc60997a94658ac4c77686cc905328783eb2b988e1446da0" exitCode=0 Sep 30 17:01:50 crc kubenswrapper[4796]: I0930 17:01:50.643749 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" event={"ID":"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83","Type":"ContainerDied","Data":"32ba7d5aa8d1798bcc60997a94658ac4c77686cc905328783eb2b988e1446da0"} Sep 30 17:01:50 crc kubenswrapper[4796]: I0930 17:01:50.683818 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-8qqvf"] Sep 30 17:01:50 crc kubenswrapper[4796]: I0930 17:01:50.696999 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rk8bx/crc-debug-8qqvf"] Sep 30 17:01:51 crc kubenswrapper[4796]: I0930 17:01:51.792347 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:51 crc kubenswrapper[4796]: I0930 17:01:51.957266 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-host\") pod \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " Sep 30 17:01:51 crc kubenswrapper[4796]: I0930 17:01:51.957366 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-host" (OuterVolumeSpecName: "host") pod "5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" (UID: "5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 17:01:51 crc kubenswrapper[4796]: I0930 17:01:51.957507 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gnqz\" (UniqueName: \"kubernetes.io/projected/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-kube-api-access-5gnqz\") pod \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\" (UID: \"5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83\") " Sep 30 17:01:51 crc kubenswrapper[4796]: I0930 17:01:51.957938 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-host\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:51 crc kubenswrapper[4796]: I0930 17:01:51.963195 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-kube-api-access-5gnqz" (OuterVolumeSpecName: "kube-api-access-5gnqz") pod "5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" (UID: "5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83"). InnerVolumeSpecName "kube-api-access-5gnqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.061324 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gnqz\" (UniqueName: \"kubernetes.io/projected/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83-kube-api-access-5gnqz\") on node \"crc\" DevicePath \"\"" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.201310 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-dlpc6_a6b170fa-6665-405f-aaa3-042e9705ed1a/kube-rbac-proxy/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.304858 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-dlpc6_a6b170fa-6665-405f-aaa3-042e9705ed1a/manager/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.452445 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-nmqln_52af31b2-ed79-46c4-840b-c2453d43a88c/kube-rbac-proxy/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.507733 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-nmqln_52af31b2-ed79-46c4-840b-c2453d43a88c/manager/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.574782 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-44kvk_24427541-c81d-4542-b20e-bdcf3a4f0e4c/kube-rbac-proxy/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.638241 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-44kvk_24427541-c81d-4542-b20e-bdcf3a4f0e4c/manager/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.685845 4796 scope.go:117] "RemoveContainer" containerID="32ba7d5aa8d1798bcc60997a94658ac4c77686cc905328783eb2b988e1446da0" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.685868 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/crc-debug-8qqvf" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.733947 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.734389 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:01:52 crc kubenswrapper[4796]: E0930 17:01:52.734613 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:01:52 crc kubenswrapper[4796]: E0930 17:01:52.734619 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.743203 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/util/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.746835 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" path="/var/lib/kubelet/pods/5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83/volumes" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.863619 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/util/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.901760 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/pull/0.log" Sep 30 17:01:52 crc kubenswrapper[4796]: I0930 17:01:52.914083 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/pull/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.088565 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/extract/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.089998 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/pull/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.096985 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fcb715ece9f005e1a03c130eeea8b9b209953c0686aefe28df3e5ad2ffczzrj_d5833467-cda7-456a-9f81-cd6b6b32a205/util/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.266991 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-nvqft_cf66eb4e-6595-40ea-b64a-b3d40e44dec9/kube-rbac-proxy/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.356625 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-nvqft_cf66eb4e-6595-40ea-b64a-b3d40e44dec9/manager/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.364508 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-4dsbl_8fba4c88-186e-4a06-8374-97e696008fcd/kube-rbac-proxy/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.451190 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-4dsbl_8fba4c88-186e-4a06-8374-97e696008fcd/manager/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.544261 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-jxwnq_70f8a96e-36a5-4824-a041-72b7ccfbe064/kube-rbac-proxy/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.562660 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-jxwnq_70f8a96e-36a5-4824-a041-72b7ccfbe064/manager/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.696792 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-97m97_8ef65c41-fdac-4065-a568-2bd9f1176adc/kube-rbac-proxy/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.833635 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-j6vsj_36b7f470-4ef5-431f-a756-04c40e1afc90/kube-rbac-proxy/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.929546 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-j6vsj_36b7f470-4ef5-431f-a756-04c40e1afc90/manager/0.log" Sep 30 17:01:53 crc kubenswrapper[4796]: I0930 17:01:53.936847 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-97m97_8ef65c41-fdac-4065-a568-2bd9f1176adc/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.063155 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-fqxdk_e50b54b7-c7c3-4614-bd2e-3b90b68e129f/kube-rbac-proxy/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.179197 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-fqxdk_e50b54b7-c7c3-4614-bd2e-3b90b68e129f/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.219910 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-j8d9f_5a8c84fa-799a-4f47-81c1-7445702e0f23/kube-rbac-proxy/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.266109 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-j8d9f_5a8c84fa-799a-4f47-81c1-7445702e0f23/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.365602 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-sb9j9_16086739-bea6-4c47-85d8-9a10497c9373/kube-rbac-proxy/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.432904 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-sb9j9_16086739-bea6-4c47-85d8-9a10497c9373/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.532515 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-mttjp_ed0b77fc-1f2a-4a53-9983-ba64a6831569/kube-rbac-proxy/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.612376 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-mttjp_ed0b77fc-1f2a-4a53-9983-ba64a6831569/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.642458 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-jj598_4131038b-1bbe-410a-9c7b-58216c527106/kube-rbac-proxy/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.791910 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-jj598_4131038b-1bbe-410a-9c7b-58216c527106/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.848732 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-q2knx_d1555f09-a29d-45bf-8bdf-5a3212d1ce0a/manager/0.log" Sep 30 17:01:54 crc kubenswrapper[4796]: I0930 17:01:54.852343 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-q2knx_d1555f09-a29d-45bf-8bdf-5a3212d1ce0a/kube-rbac-proxy/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.010279 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-p2jwc_df653968-5944-4e8d-9e82-7e19a539d997/kube-rbac-proxy/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.039604 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-p2jwc_df653968-5944-4e8d-9e82-7e19a539d997/manager/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.180660 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55476bd9c7-gvldb_d9b3d318-c9a7-412a-a926-6f840dd14df6/kube-rbac-proxy/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.275524 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7ff58777bc-cfdvx_107cc98e-9b37-4a5f-99f3-73350360e2ab/kube-rbac-proxy/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.587952 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7ff58777bc-cfdvx_107cc98e-9b37-4a5f-99f3-73350360e2ab/operator/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.590172 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f5fb8_3d4983f8-cba0-43ba-b756-0cab71ed31f8/registry-server/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.733013 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:01:55 crc kubenswrapper[4796]: E0930 17:01:55.733667 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.744143 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-jqbdb_5cadf327-7ef7-439e-b692-5eef55c1b666/kube-rbac-proxy/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.883937 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-jqbdb_5cadf327-7ef7-439e-b692-5eef55c1b666/manager/0.log" Sep 30 17:01:55 crc kubenswrapper[4796]: I0930 17:01:55.968170 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-whxjg_ae3399ed-5451-44ba-9872-73d709faab31/kube-rbac-proxy/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.072264 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-whxjg_ae3399ed-5451-44ba-9872-73d709faab31/manager/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.128395 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-dwv6m_d70c88cd-5ff3-49d9-b581-54860881ea39/operator/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.310168 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-f45lz_cb7865b9-6ef2-4cc1-92e5-0753a28e43c1/kube-rbac-proxy/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.397451 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-f45lz_cb7865b9-6ef2-4cc1-92e5-0753a28e43c1/manager/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.403450 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55476bd9c7-gvldb_d9b3d318-c9a7-412a-a926-6f840dd14df6/manager/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.456416 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6db74cfd6d-l8qfg_530de9cb-0bdc-4773-a4ff-29b992d1942f/kube-rbac-proxy/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.622297 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-dntbx_40fc5188-af1b-454f-b7f4-6c1ae1c4cf71/manager/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.645197 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-dntbx_40fc5188-af1b-454f-b7f4-6c1ae1c4cf71/kube-rbac-proxy/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.660765 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6db74cfd6d-l8qfg_530de9cb-0bdc-4773-a4ff-29b992d1942f/manager/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.734451 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:01:56 crc kubenswrapper[4796]: E0930 17:01:56.734706 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.794474 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-j4ntp_2a580ec9-b0df-41bb-a272-cc4076cacdba/kube-rbac-proxy/0.log" Sep 30 17:01:56 crc kubenswrapper[4796]: I0930 17:01:56.826824 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-j4ntp_2a580ec9-b0df-41bb-a272-cc4076cacdba/manager/0.log" Sep 30 17:02:05 crc kubenswrapper[4796]: I0930 17:02:05.733959 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:02:05 crc kubenswrapper[4796]: E0930 17:02:05.734803 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:02:06 crc kubenswrapper[4796]: I0930 17:02:06.734421 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:02:06 crc kubenswrapper[4796]: E0930 17:02:06.735276 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:02:07 crc kubenswrapper[4796]: I0930 17:02:07.734254 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:02:07 crc kubenswrapper[4796]: E0930 17:02:07.734581 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:02:11 crc kubenswrapper[4796]: I0930 17:02:11.502388 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-cwrnl_be7f735b-ccce-4060-915d-58aea627d7f5/control-plane-machine-set-operator/0.log" Sep 30 17:02:11 crc kubenswrapper[4796]: I0930 17:02:11.681889 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pbnlg_ecf5a09b-2113-4202-83c1-ad8c77520791/kube-rbac-proxy/0.log" Sep 30 17:02:11 crc kubenswrapper[4796]: I0930 17:02:11.700889 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pbnlg_ecf5a09b-2113-4202-83c1-ad8c77520791/machine-api-operator/0.log" Sep 30 17:02:11 crc kubenswrapper[4796]: I0930 17:02:11.733322 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:02:11 crc kubenswrapper[4796]: E0930 17:02:11.733786 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:02:18 crc kubenswrapper[4796]: I0930 17:02:18.739917 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:02:18 crc kubenswrapper[4796]: E0930 17:02:18.740711 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:02:19 crc kubenswrapper[4796]: I0930 17:02:19.734468 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:02:19 crc kubenswrapper[4796]: E0930 17:02:19.735556 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:02:22 crc kubenswrapper[4796]: I0930 17:02:22.733026 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:02:22 crc kubenswrapper[4796]: E0930 17:02:22.733752 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:02:23 crc kubenswrapper[4796]: I0930 17:02:23.733641 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:02:23 crc kubenswrapper[4796]: E0930 17:02:23.735054 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:02:23 crc kubenswrapper[4796]: I0930 17:02:23.851473 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-s2jmq_8245bb2a-a295-4493-b9fc-88469011a23c/cert-manager-controller/0.log" Sep 30 17:02:24 crc kubenswrapper[4796]: I0930 17:02:24.007558 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-zl4xg_4bd447e5-097a-4b02-be6e-6af942a45369/cert-manager-cainjector/0.log" Sep 30 17:02:24 crc kubenswrapper[4796]: I0930 17:02:24.053959 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hgxtd_df8942ab-8a15-4023-a58f-c36972391914/cert-manager-webhook/0.log" Sep 30 17:02:30 crc kubenswrapper[4796]: I0930 17:02:30.733861 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:02:30 crc kubenswrapper[4796]: E0930 17:02:30.734718 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:02:33 crc kubenswrapper[4796]: I0930 17:02:33.734220 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:02:33 crc kubenswrapper[4796]: E0930 17:02:33.734998 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:02:36 crc kubenswrapper[4796]: I0930 17:02:36.399196 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-96z6n_55affe5c-1791-49c1-9218-cbc57c6ee847/nmstate-console-plugin/0.log" Sep 30 17:02:36 crc kubenswrapper[4796]: I0930 17:02:36.584445 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hpgrl_bf0cf03b-b835-403a-83c5-dd587a330d94/nmstate-handler/0.log" Sep 30 17:02:36 crc kubenswrapper[4796]: I0930 17:02:36.620707 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-9fgrw_541be51b-57e9-42d8-b0f7-02a1276c910f/kube-rbac-proxy/0.log" Sep 30 17:02:36 crc kubenswrapper[4796]: I0930 17:02:36.655887 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-9fgrw_541be51b-57e9-42d8-b0f7-02a1276c910f/nmstate-metrics/0.log" Sep 30 17:02:36 crc kubenswrapper[4796]: I0930 17:02:36.802484 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-jpgzs_70baf97d-1c3a-4106-9754-9369d381faf1/nmstate-operator/0.log" Sep 30 17:02:36 crc kubenswrapper[4796]: I0930 17:02:36.876365 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-9xkld_90f4fdf8-d73e-47f1-9e33-41f58baf256b/nmstate-webhook/0.log" Sep 30 17:02:37 crc kubenswrapper[4796]: I0930 17:02:37.733963 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:02:37 crc kubenswrapper[4796]: I0930 17:02:37.734207 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:02:37 crc kubenswrapper[4796]: E0930 17:02:37.734250 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:02:37 crc kubenswrapper[4796]: E0930 17:02:37.734456 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:02:45 crc kubenswrapper[4796]: I0930 17:02:45.732864 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:02:45 crc kubenswrapper[4796]: I0930 17:02:45.733384 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:02:45 crc kubenswrapper[4796]: E0930 17:02:45.733697 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:02:45 crc kubenswrapper[4796]: E0930 17:02:45.733696 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:02:48 crc kubenswrapper[4796]: I0930 17:02:48.721560 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-769bbf7ffd-fzdwf_233c52ec-0b90-41d5-9442-bb0ce913adaf/kube-rbac-proxy/0.log" Sep 30 17:02:48 crc kubenswrapper[4796]: I0930 17:02:48.741854 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-769bbf7ffd-fzdwf_233c52ec-0b90-41d5-9442-bb0ce913adaf/manager/0.log" Sep 30 17:02:50 crc kubenswrapper[4796]: I0930 17:02:50.733020 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:02:50 crc kubenswrapper[4796]: E0930 17:02:50.733540 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:02:52 crc kubenswrapper[4796]: I0930 17:02:52.733604 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:02:52 crc kubenswrapper[4796]: E0930 17:02:52.734469 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:02:56 crc kubenswrapper[4796]: I0930 17:02:56.733682 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:02:56 crc kubenswrapper[4796]: E0930 17:02:56.734523 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:03:00 crc kubenswrapper[4796]: I0930 17:03:00.733683 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:03:00 crc kubenswrapper[4796]: E0930 17:03:00.734227 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:03:01 crc kubenswrapper[4796]: I0930 17:03:01.010287 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-fcc886d58-q8fnc_ae187c99-f8dd-4c13-a001-3cb453b62739/cluster-logging-operator/0.log" Sep 30 17:03:01 crc kubenswrapper[4796]: I0930 17:03:01.171895 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-l7td6_da982abe-554b-4964-ae74-ec0b4fce349d/collector/0.log" Sep 30 17:03:01 crc kubenswrapper[4796]: I0930 17:03:01.734077 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:03:01 crc kubenswrapper[4796]: E0930 17:03:01.734676 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:03:03 crc kubenswrapper[4796]: I0930 17:03:03.733780 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:03:03 crc kubenswrapper[4796]: E0930 17:03:03.734367 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:03:09 crc kubenswrapper[4796]: I0930 17:03:09.733838 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:03:10 crc kubenswrapper[4796]: I0930 17:03:10.525441 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41"} Sep 30 17:03:12 crc kubenswrapper[4796]: I0930 17:03:12.732997 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:03:12 crc kubenswrapper[4796]: E0930 17:03:12.733624 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.341170 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-z8rjn_7b713269-c9b4-4619-9f57-7b932bcc2977/kube-rbac-proxy/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.417824 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-z8rjn_7b713269-c9b4-4619-9f57-7b932bcc2977/controller/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.497329 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-frr-files/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.729669 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-metrics/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.731169 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-reloader/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.734115 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:03:13 crc kubenswrapper[4796]: E0930 17:03:13.734436 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.748300 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-reloader/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.752929 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-frr-files/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.962937 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-frr-files/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.968220 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-reloader/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.979325 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-metrics/0.log" Sep 30 17:03:13 crc kubenswrapper[4796]: I0930 17:03:13.984970 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-metrics/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.121758 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-frr-files/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.128843 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-metrics/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.145777 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/cp-reloader/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.169545 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/controller/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.327609 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/frr-metrics/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.337086 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/kube-rbac-proxy/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.411884 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/kube-rbac-proxy-frr/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.604943 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:03:14 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:03:14 crc kubenswrapper[4796]: > Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.631412 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/reloader/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.632931 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-bq9nr_d03550b5-47a9-41ae-ad9b-1cc7d57c3335/frr-k8s-webhook-server/0.log" Sep 30 17:03:14 crc kubenswrapper[4796]: I0930 17:03:14.840910 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-588f64bf97-2tdzp_184ca539-6770-44bf-8461-bbde8fce05e1/manager/0.log" Sep 30 17:03:15 crc kubenswrapper[4796]: I0930 17:03:15.022941 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-54894f57bf-kfz9w_81e041f0-90d3-44a2-b3bd-d15212a55e24/webhook-server/0.log" Sep 30 17:03:15 crc kubenswrapper[4796]: I0930 17:03:15.126955 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tl8x6_5dd558be-ca13-43fc-9548-0ec92e3a2b69/kube-rbac-proxy/0.log" Sep 30 17:03:15 crc kubenswrapper[4796]: I0930 17:03:15.777722 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2psf_d09a2502-5fca-4442-9f3b-f245b82ffd77/frr/0.log" Sep 30 17:03:15 crc kubenswrapper[4796]: I0930 17:03:15.814774 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tl8x6_5dd558be-ca13-43fc-9548-0ec92e3a2b69/speaker/0.log" Sep 30 17:03:17 crc kubenswrapper[4796]: I0930 17:03:17.614402 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:03:17 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:03:17 crc kubenswrapper[4796]: > Sep 30 17:03:17 crc kubenswrapper[4796]: I0930 17:03:17.732880 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:03:17 crc kubenswrapper[4796]: E0930 17:03:17.733211 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:03:20 crc kubenswrapper[4796]: I0930 17:03:20.654418 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:03:20 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:03:20 crc kubenswrapper[4796]: > Sep 30 17:03:20 crc kubenswrapper[4796]: I0930 17:03:20.655121 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:03:20 crc kubenswrapper[4796]: I0930 17:03:20.655972 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:03:20 crc kubenswrapper[4796]: I0930 17:03:20.656050 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" gracePeriod=30 Sep 30 17:03:21 crc kubenswrapper[4796]: E0930 17:03:21.077085 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:03:21 crc kubenswrapper[4796]: I0930 17:03:21.661231 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" exitCode=0 Sep 30 17:03:21 crc kubenswrapper[4796]: I0930 17:03:21.661318 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41"} Sep 30 17:03:21 crc kubenswrapper[4796]: I0930 17:03:21.661596 4796 scope.go:117] "RemoveContainer" containerID="31ece02dfdd6ea2a3c8b6acb3038c8bc7154b10e72ba89f7e8f6e49dcf6279e4" Sep 30 17:03:21 crc kubenswrapper[4796]: I0930 17:03:21.662513 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:03:21 crc kubenswrapper[4796]: E0930 17:03:21.662924 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:03:24 crc kubenswrapper[4796]: I0930 17:03:24.734076 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:03:24 crc kubenswrapper[4796]: I0930 17:03:24.734735 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:03:24 crc kubenswrapper[4796]: E0930 17:03:24.734798 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:03:24 crc kubenswrapper[4796]: E0930 17:03:24.734966 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:03:28 crc kubenswrapper[4796]: I0930 17:03:28.807422 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/util/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.064501 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/util/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.078834 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/pull/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.149189 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/pull/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.366873 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/pull/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.374421 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/util/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.374543 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694xvp6n_5368fc26-bf5c-4c4d-add0-46d7f1fc0346/extract/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.520882 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/util/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.723769 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/pull/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.774434 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/pull/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.795074 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/util/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.949334 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/util/0.log" Sep 30 17:03:29 crc kubenswrapper[4796]: I0930 17:03:29.984395 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/extract/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.041326 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bckf8p2_9a961850-d9b5-4e08-bbc6-cc2302da4a45/pull/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.163965 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/util/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.351465 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/util/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.360635 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/pull/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.377746 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/pull/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.568763 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/pull/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.610326 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/extract/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.625990 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dptwsq_0d843fc8-0ae6-4498-97b4-9b40bafd59d0/util/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.777666 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/util/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.953678 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/pull/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.966040 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/util/0.log" Sep 30 17:03:30 crc kubenswrapper[4796]: I0930 17:03:30.997409 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/pull/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.189720 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/extract/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.203383 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/util/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.241176 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0cg2s7_f910fbb3-44e8-4db1-a65b-5fb0297676bf/pull/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.403553 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/extract-utilities/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.571768 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/extract-utilities/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.597056 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/extract-content/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.624189 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/extract-content/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.825481 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/extract-content/0.log" Sep 30 17:03:31 crc kubenswrapper[4796]: I0930 17:03:31.825580 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/extract-utilities/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.085224 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/extract-utilities/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.232155 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/extract-utilities/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.275227 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/extract-content/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.315603 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/extract-content/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.369461 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zw4ct_0723eea3-9e01-4945-be27-a9bbfd46cb3c/registry-server/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.483504 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/extract-utilities/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.513262 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/extract-content/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.563591 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/util/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.734132 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:03:32 crc kubenswrapper[4796]: E0930 17:03:32.734636 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.793276 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/util/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.812925 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/pull/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.866205 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/pull/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.985198 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xln5d_1d2e366d-8e88-4f1b-b2da-8db4ee157808/registry-server/0.log" Sep 30 17:03:32 crc kubenswrapper[4796]: I0930 17:03:32.994026 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/pull/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.036902 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/extract/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.053194 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170lz4fj_c5876a04-724a-4dda-a51d-d84db778caf8/util/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.158317 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/util/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.327584 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/util/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.371679 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/pull/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.391486 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/pull/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.599731 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/util/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.603179 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/extract/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.607919 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d965n7wd_1aaa2ce2-f1ef-449b-a3e8-82bf5f8cbc64/pull/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.608891 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-lchcj_f6ab27c4-6027-4d97-88c0-a8ad5eac035a/marketplace-operator/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.771057 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/extract-utilities/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.925028 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/extract-content/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.943382 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/extract-utilities/0.log" Sep 30 17:03:33 crc kubenswrapper[4796]: I0930 17:03:33.948205 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/extract-content/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.138549 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/extract-utilities/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.159558 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/extract-utilities/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.182236 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/extract-content/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.293629 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n8r5j_399cf9ef-2274-4e40-bf07-f279f6d3730c/registry-server/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.389554 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/extract-utilities/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.390251 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/extract-content/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.463144 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/extract-content/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.559602 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/extract-utilities/0.log" Sep 30 17:03:34 crc kubenswrapper[4796]: I0930 17:03:34.560665 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/extract-content/0.log" Sep 30 17:03:35 crc kubenswrapper[4796]: I0930 17:03:35.187033 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n28l7_227b1cd2-3c9b-44c2-baa9-e29119fe40de/registry-server/0.log" Sep 30 17:03:35 crc kubenswrapper[4796]: I0930 17:03:35.732659 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:03:35 crc kubenswrapper[4796]: E0930 17:03:35.732921 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:03:36 crc kubenswrapper[4796]: I0930 17:03:36.734292 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:03:36 crc kubenswrapper[4796]: E0930 17:03:36.734854 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:03:39 crc kubenswrapper[4796]: I0930 17:03:39.733207 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:03:39 crc kubenswrapper[4796]: E0930 17:03:39.733911 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:03:45 crc kubenswrapper[4796]: I0930 17:03:45.733110 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:03:45 crc kubenswrapper[4796]: E0930 17:03:45.733810 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:03:46 crc kubenswrapper[4796]: I0930 17:03:46.722479 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-l7klx_c18ee88c-1de9-4356-95bd-6df7da966490/prometheus-operator/0.log" Sep 30 17:03:46 crc kubenswrapper[4796]: I0930 17:03:46.864085 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6fbbddfb9d-hrh9z_a9a06303-d13f-4a67-b856-95e577e0fab7/prometheus-operator-admission-webhook/0.log" Sep 30 17:03:46 crc kubenswrapper[4796]: I0930 17:03:46.884200 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6fbbddfb9d-tsdj6_f1277eb0-bc20-4e0e-a3f8-7a25cbf30168/prometheus-operator-admission-webhook/0.log" Sep 30 17:03:47 crc kubenswrapper[4796]: I0930 17:03:47.015250 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-929vt_0454eedd-98d0-449a-ab54-ceb594db9c5e/operator/0.log" Sep 30 17:03:47 crc kubenswrapper[4796]: I0930 17:03:47.079955 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-xvbbv_1e62a431-bc1c-4624-8463-93a4f9b7b32e/perses-operator/0.log" Sep 30 17:03:49 crc kubenswrapper[4796]: I0930 17:03:49.734221 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:03:49 crc kubenswrapper[4796]: E0930 17:03:49.735690 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:03:50 crc kubenswrapper[4796]: I0930 17:03:50.734143 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:03:50 crc kubenswrapper[4796]: I0930 17:03:50.734622 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:03:50 crc kubenswrapper[4796]: E0930 17:03:50.734851 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:03:50 crc kubenswrapper[4796]: E0930 17:03:50.734931 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8bbws_openshift-machine-config-operator(670c655e-3953-4773-84ef-19c678d482f9)\"" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" Sep 30 17:03:57 crc kubenswrapper[4796]: I0930 17:03:57.733840 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:03:57 crc kubenswrapper[4796]: E0930 17:03:57.736169 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:03:59 crc kubenswrapper[4796]: I0930 17:03:59.563218 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-769bbf7ffd-fzdwf_233c52ec-0b90-41d5-9442-bb0ce913adaf/manager/0.log" Sep 30 17:03:59 crc kubenswrapper[4796]: I0930 17:03:59.612107 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-769bbf7ffd-fzdwf_233c52ec-0b90-41d5-9442-bb0ce913adaf/kube-rbac-proxy/0.log" Sep 30 17:04:02 crc kubenswrapper[4796]: I0930 17:04:02.733706 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:04:03 crc kubenswrapper[4796]: I0930 17:04:03.092301 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"0e16983f6a64c78a8ee2d7ba623c69e9a76d02d91fe4467bdd766d749b98a1d0"} Sep 30 17:04:04 crc kubenswrapper[4796]: I0930 17:04:04.737001 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:04:04 crc kubenswrapper[4796]: E0930 17:04:04.737449 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:04:04 crc kubenswrapper[4796]: I0930 17:04:04.737852 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:04:04 crc kubenswrapper[4796]: E0930 17:04:04.738049 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:04:10 crc kubenswrapper[4796]: I0930 17:04:10.734701 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:04:11 crc kubenswrapper[4796]: I0930 17:04:11.178408 4796 generic.go:334] "Generic (PLEG): container finished" podID="15e69566-5b25-4929-a40b-1a3e2d710800" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" exitCode=2 Sep 30 17:04:11 crc kubenswrapper[4796]: I0930 17:04:11.178485 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" event={"ID":"15e69566-5b25-4929-a40b-1a3e2d710800","Type":"ContainerDied","Data":"e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59"} Sep 30 17:04:11 crc kubenswrapper[4796]: I0930 17:04:11.178708 4796 scope.go:117] "RemoveContainer" containerID="ee20bf5c4f7c18edd7db98b8bfda58b23a8bbde06fd073ec63a2edd3e71659d1" Sep 30 17:04:11 crc kubenswrapper[4796]: I0930 17:04:11.179410 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:04:11 crc kubenswrapper[4796]: E0930 17:04:11.179722 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:04:15 crc kubenswrapper[4796]: I0930 17:04:15.797411 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 17:04:15 crc kubenswrapper[4796]: I0930 17:04:15.798811 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:04:15 crc kubenswrapper[4796]: E0930 17:04:15.799192 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:04:17 crc kubenswrapper[4796]: I0930 17:04:17.733608 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:04:17 crc kubenswrapper[4796]: E0930 17:04:17.734068 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:04:19 crc kubenswrapper[4796]: I0930 17:04:19.733449 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:04:20 crc kubenswrapper[4796]: E0930 17:04:20.264247 4796 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.38:39684->38.102.83.38:37569: write tcp 38.102.83.38:39684->38.102.83.38:37569: write: broken pipe Sep 30 17:04:20 crc kubenswrapper[4796]: I0930 17:04:20.281836 4796 generic.go:334] "Generic (PLEG): container finished" podID="165eb10b-4c60-4635-928e-c04a72e9b4f5" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" exitCode=2 Sep 30 17:04:20 crc kubenswrapper[4796]: I0930 17:04:20.281885 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" event={"ID":"165eb10b-4c60-4635-928e-c04a72e9b4f5","Type":"ContainerDied","Data":"60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a"} Sep 30 17:04:20 crc kubenswrapper[4796]: I0930 17:04:20.281925 4796 scope.go:117] "RemoveContainer" containerID="fabce9cb101b9ce95de726cef79ee82ad70b533ab71bca888b2cb9fd015bc641" Sep 30 17:04:20 crc kubenswrapper[4796]: I0930 17:04:20.282708 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:04:20 crc kubenswrapper[4796]: E0930 17:04:20.283093 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:04:21 crc kubenswrapper[4796]: I0930 17:04:21.375158 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 17:04:21 crc kubenswrapper[4796]: I0930 17:04:21.376232 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:04:21 crc kubenswrapper[4796]: E0930 17:04:21.376504 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:04:28 crc kubenswrapper[4796]: I0930 17:04:28.735721 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:04:28 crc kubenswrapper[4796]: E0930 17:04:28.736417 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:04:28 crc kubenswrapper[4796]: I0930 17:04:28.743402 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:04:28 crc kubenswrapper[4796]: E0930 17:04:28.743629 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:04:36 crc kubenswrapper[4796]: I0930 17:04:36.735001 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:04:36 crc kubenswrapper[4796]: E0930 17:04:36.735803 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:04:40 crc kubenswrapper[4796]: I0930 17:04:40.797239 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" Sep 30 17:04:40 crc kubenswrapper[4796]: I0930 17:04:40.798446 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:04:40 crc kubenswrapper[4796]: E0930 17:04:40.798773 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:04:41 crc kubenswrapper[4796]: I0930 17:04:41.375354 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" Sep 30 17:04:41 crc kubenswrapper[4796]: I0930 17:04:41.376038 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:04:41 crc kubenswrapper[4796]: E0930 17:04:41.376472 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:04:41 crc kubenswrapper[4796]: I0930 17:04:41.736173 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:04:41 crc kubenswrapper[4796]: E0930 17:04:41.737029 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:04:51 crc kubenswrapper[4796]: I0930 17:04:51.733835 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:04:51 crc kubenswrapper[4796]: E0930 17:04:51.734628 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:04:54 crc kubenswrapper[4796]: I0930 17:04:54.735769 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:04:54 crc kubenswrapper[4796]: E0930 17:04:54.736709 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:04:55 crc kubenswrapper[4796]: I0930 17:04:55.733392 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:04:55 crc kubenswrapper[4796]: E0930 17:04:55.733866 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.040611 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pxxrz"] Sep 30 17:05:00 crc kubenswrapper[4796]: E0930 17:05:00.041768 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" containerName="container-00" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.041786 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" containerName="container-00" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.042110 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5db2d5c0-ac0b-4d7a-bc88-ff7d53bc3e83" containerName="container-00" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.044052 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.051074 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pxxrz"] Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.093437 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-utilities\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.093706 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw69c\" (UniqueName: \"kubernetes.io/projected/8d104c70-9f39-4075-82a0-d6102d512746-kube-api-access-pw69c\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.093742 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-catalog-content\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.195359 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw69c\" (UniqueName: \"kubernetes.io/projected/8d104c70-9f39-4075-82a0-d6102d512746-kube-api-access-pw69c\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.195766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-catalog-content\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.196063 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-utilities\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.196537 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-catalog-content\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.196954 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-utilities\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.235495 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw69c\" (UniqueName: \"kubernetes.io/projected/8d104c70-9f39-4075-82a0-d6102d512746-kube-api-access-pw69c\") pod \"certified-operators-pxxrz\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.387049 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:00 crc kubenswrapper[4796]: I0930 17:05:00.893739 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pxxrz"] Sep 30 17:05:01 crc kubenswrapper[4796]: I0930 17:05:01.893224 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d104c70-9f39-4075-82a0-d6102d512746" containerID="a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10" exitCode=0 Sep 30 17:05:01 crc kubenswrapper[4796]: I0930 17:05:01.893296 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerDied","Data":"a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10"} Sep 30 17:05:01 crc kubenswrapper[4796]: I0930 17:05:01.893610 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerStarted","Data":"1195ca2a6896f64466a0ea4c3235f3c58491e5c91cefe56fe823af67711186d1"} Sep 30 17:05:01 crc kubenswrapper[4796]: I0930 17:05:01.895704 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 17:05:03 crc kubenswrapper[4796]: I0930 17:05:03.915829 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerStarted","Data":"fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0"} Sep 30 17:05:04 crc kubenswrapper[4796]: E0930 17:05:04.918882 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d104c70_9f39_4075_82a0_d6102d512746.slice/crio-conmon-fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d104c70_9f39_4075_82a0_d6102d512746.slice/crio-fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0.scope\": RecentStats: unable to find data in memory cache]" Sep 30 17:05:04 crc kubenswrapper[4796]: I0930 17:05:04.933820 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d104c70-9f39-4075-82a0-d6102d512746" containerID="fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0" exitCode=0 Sep 30 17:05:04 crc kubenswrapper[4796]: I0930 17:05:04.933891 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerDied","Data":"fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0"} Sep 30 17:05:05 crc kubenswrapper[4796]: I0930 17:05:05.734270 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:05:05 crc kubenswrapper[4796]: E0930 17:05:05.735056 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:05:06 crc kubenswrapper[4796]: I0930 17:05:06.964232 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerStarted","Data":"cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df"} Sep 30 17:05:07 crc kubenswrapper[4796]: I0930 17:05:07.005663 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pxxrz" podStartSLOduration=3.274010395 podStartE2EDuration="7.005640577s" podCreationTimestamp="2025-09-30 17:05:00 +0000 UTC" firstStartedPulling="2025-09-30 17:05:01.895459354 +0000 UTC m=+3193.908737881" lastFinishedPulling="2025-09-30 17:05:05.627089506 +0000 UTC m=+3197.640368063" observedRunningTime="2025-09-30 17:05:06.992310542 +0000 UTC m=+3199.005589069" watchObservedRunningTime="2025-09-30 17:05:07.005640577 +0000 UTC m=+3199.018919114" Sep 30 17:05:07 crc kubenswrapper[4796]: I0930 17:05:07.733375 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:05:07 crc kubenswrapper[4796]: E0930 17:05:07.734188 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:05:10 crc kubenswrapper[4796]: I0930 17:05:10.387455 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:10 crc kubenswrapper[4796]: I0930 17:05:10.387837 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:10 crc kubenswrapper[4796]: I0930 17:05:10.453531 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:10 crc kubenswrapper[4796]: I0930 17:05:10.734479 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:05:10 crc kubenswrapper[4796]: E0930 17:05:10.734929 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:05:11 crc kubenswrapper[4796]: I0930 17:05:11.075370 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:11 crc kubenswrapper[4796]: I0930 17:05:11.143610 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pxxrz"] Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.036182 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pxxrz" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="registry-server" containerID="cri-o://cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df" gracePeriod=2 Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.530838 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.572904 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-utilities\") pod \"8d104c70-9f39-4075-82a0-d6102d512746\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.574077 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-catalog-content\") pod \"8d104c70-9f39-4075-82a0-d6102d512746\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.574239 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw69c\" (UniqueName: \"kubernetes.io/projected/8d104c70-9f39-4075-82a0-d6102d512746-kube-api-access-pw69c\") pod \"8d104c70-9f39-4075-82a0-d6102d512746\" (UID: \"8d104c70-9f39-4075-82a0-d6102d512746\") " Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.573847 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-utilities" (OuterVolumeSpecName: "utilities") pod "8d104c70-9f39-4075-82a0-d6102d512746" (UID: "8d104c70-9f39-4075-82a0-d6102d512746"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.602310 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d104c70-9f39-4075-82a0-d6102d512746-kube-api-access-pw69c" (OuterVolumeSpecName: "kube-api-access-pw69c") pod "8d104c70-9f39-4075-82a0-d6102d512746" (UID: "8d104c70-9f39-4075-82a0-d6102d512746"). InnerVolumeSpecName "kube-api-access-pw69c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.628712 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d104c70-9f39-4075-82a0-d6102d512746" (UID: "8d104c70-9f39-4075-82a0-d6102d512746"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.676475 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.676504 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d104c70-9f39-4075-82a0-d6102d512746-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 17:05:13 crc kubenswrapper[4796]: I0930 17:05:13.676514 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw69c\" (UniqueName: \"kubernetes.io/projected/8d104c70-9f39-4075-82a0-d6102d512746-kube-api-access-pw69c\") on node \"crc\" DevicePath \"\"" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.062505 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d104c70-9f39-4075-82a0-d6102d512746" containerID="cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df" exitCode=0 Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.062587 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerDied","Data":"cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df"} Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.062645 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pxxrz" event={"ID":"8d104c70-9f39-4075-82a0-d6102d512746","Type":"ContainerDied","Data":"1195ca2a6896f64466a0ea4c3235f3c58491e5c91cefe56fe823af67711186d1"} Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.062680 4796 scope.go:117] "RemoveContainer" containerID="cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.062909 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pxxrz" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.091695 4796 scope.go:117] "RemoveContainer" containerID="fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.152495 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pxxrz"] Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.152814 4796 scope.go:117] "RemoveContainer" containerID="a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.162064 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pxxrz"] Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.193397 4796 scope.go:117] "RemoveContainer" containerID="cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df" Sep 30 17:05:14 crc kubenswrapper[4796]: E0930 17:05:14.193920 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df\": container with ID starting with cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df not found: ID does not exist" containerID="cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.194167 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df"} err="failed to get container status \"cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df\": rpc error: code = NotFound desc = could not find container \"cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df\": container with ID starting with cf5e344a650678fd4617bc7f17c2f9b4e45bcc38051716238e4acb623f4587df not found: ID does not exist" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.194230 4796 scope.go:117] "RemoveContainer" containerID="fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0" Sep 30 17:05:14 crc kubenswrapper[4796]: E0930 17:05:14.194814 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0\": container with ID starting with fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0 not found: ID does not exist" containerID="fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.194859 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0"} err="failed to get container status \"fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0\": rpc error: code = NotFound desc = could not find container \"fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0\": container with ID starting with fb1476bd54c8ec2cfcd1d6bdbc48e8b2659d54ffa0e6ea500bb3a36a854898b0 not found: ID does not exist" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.194888 4796 scope.go:117] "RemoveContainer" containerID="a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10" Sep 30 17:05:14 crc kubenswrapper[4796]: E0930 17:05:14.195372 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10\": container with ID starting with a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10 not found: ID does not exist" containerID="a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.195414 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10"} err="failed to get container status \"a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10\": rpc error: code = NotFound desc = could not find container \"a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10\": container with ID starting with a373a6bc12661b8d1975d69e5e0fd3b84fb13968e37ca447ba86f220544bbe10 not found: ID does not exist" Sep 30 17:05:14 crc kubenswrapper[4796]: I0930 17:05:14.754463 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d104c70-9f39-4075-82a0-d6102d512746" path="/var/lib/kubelet/pods/8d104c70-9f39-4075-82a0-d6102d512746/volumes" Sep 30 17:05:17 crc kubenswrapper[4796]: I0930 17:05:17.733676 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:05:17 crc kubenswrapper[4796]: E0930 17:05:17.737417 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:05:20 crc kubenswrapper[4796]: I0930 17:05:20.733683 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:05:20 crc kubenswrapper[4796]: E0930 17:05:20.734336 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:05:24 crc kubenswrapper[4796]: I0930 17:05:24.734725 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:05:24 crc kubenswrapper[4796]: E0930 17:05:24.735590 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:05:30 crc kubenswrapper[4796]: I0930 17:05:30.734013 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:05:30 crc kubenswrapper[4796]: E0930 17:05:30.735036 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:05:35 crc kubenswrapper[4796]: I0930 17:05:35.733341 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:05:35 crc kubenswrapper[4796]: E0930 17:05:35.735228 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:05:39 crc kubenswrapper[4796]: I0930 17:05:39.739468 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:05:39 crc kubenswrapper[4796]: E0930 17:05:39.741691 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:05:43 crc kubenswrapper[4796]: I0930 17:05:43.733175 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:05:43 crc kubenswrapper[4796]: E0930 17:05:43.734177 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:05:46 crc kubenswrapper[4796]: I0930 17:05:46.734113 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:05:46 crc kubenswrapper[4796]: E0930 17:05:46.734718 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:05:51 crc kubenswrapper[4796]: I0930 17:05:51.558827 4796 generic.go:334] "Generic (PLEG): container finished" podID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerID="c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35" exitCode=0 Sep 30 17:05:51 crc kubenswrapper[4796]: I0930 17:05:51.558954 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" event={"ID":"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58","Type":"ContainerDied","Data":"c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35"} Sep 30 17:05:51 crc kubenswrapper[4796]: I0930 17:05:51.560017 4796 scope.go:117] "RemoveContainer" containerID="c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35" Sep 30 17:05:51 crc kubenswrapper[4796]: I0930 17:05:51.667648 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rk8bx_must-gather-xkh7b_90a2b4ab-61f5-49b9-9be8-aeab05d1ba58/gather/0.log" Sep 30 17:05:51 crc kubenswrapper[4796]: I0930 17:05:51.732900 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:05:51 crc kubenswrapper[4796]: E0930 17:05:51.733248 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:05:55 crc kubenswrapper[4796]: I0930 17:05:55.734004 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:05:55 crc kubenswrapper[4796]: E0930 17:05:55.736008 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:05:59 crc kubenswrapper[4796]: I0930 17:05:59.735768 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:05:59 crc kubenswrapper[4796]: E0930 17:05:59.736727 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.064454 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rk8bx/must-gather-xkh7b"] Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.064824 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="copy" containerID="cri-o://b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe" gracePeriod=2 Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.097099 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rk8bx/must-gather-xkh7b"] Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.564101 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rk8bx_must-gather-xkh7b_90a2b4ab-61f5-49b9-9be8-aeab05d1ba58/copy/0.log" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.565125 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.684779 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rk8bx_must-gather-xkh7b_90a2b4ab-61f5-49b9-9be8-aeab05d1ba58/copy/0.log" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.687181 4796 generic.go:334] "Generic (PLEG): container finished" podID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerID="b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe" exitCode=143 Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.687239 4796 scope.go:117] "RemoveContainer" containerID="b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.687459 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rk8bx/must-gather-xkh7b" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.701384 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fnsb\" (UniqueName: \"kubernetes.io/projected/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-kube-api-access-9fnsb\") pod \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.701813 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-must-gather-output\") pod \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\" (UID: \"90a2b4ab-61f5-49b9-9be8-aeab05d1ba58\") " Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.708578 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-kube-api-access-9fnsb" (OuterVolumeSpecName: "kube-api-access-9fnsb") pod "90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" (UID: "90a2b4ab-61f5-49b9-9be8-aeab05d1ba58"). InnerVolumeSpecName "kube-api-access-9fnsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.730206 4796 scope.go:117] "RemoveContainer" containerID="c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.809450 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fnsb\" (UniqueName: \"kubernetes.io/projected/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-kube-api-access-9fnsb\") on node \"crc\" DevicePath \"\"" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.825660 4796 scope.go:117] "RemoveContainer" containerID="b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe" Sep 30 17:06:00 crc kubenswrapper[4796]: E0930 17:06:00.826551 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe\": container with ID starting with b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe not found: ID does not exist" containerID="b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.826591 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe"} err="failed to get container status \"b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe\": rpc error: code = NotFound desc = could not find container \"b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe\": container with ID starting with b7d71a696a3878a6c97d7f1e8b82ceca5cfdc1e13da469f0b7f680ff3af38efe not found: ID does not exist" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.826617 4796 scope.go:117] "RemoveContainer" containerID="c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35" Sep 30 17:06:00 crc kubenswrapper[4796]: E0930 17:06:00.827111 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35\": container with ID starting with c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35 not found: ID does not exist" containerID="c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.827170 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35"} err="failed to get container status \"c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35\": rpc error: code = NotFound desc = could not find container \"c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35\": container with ID starting with c2504078cf0ed85af4ce6254c3bbeeeacb838428c30c13f1c8064f1542f5be35 not found: ID does not exist" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.906474 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" (UID: "90a2b4ab-61f5-49b9-9be8-aeab05d1ba58"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:06:00 crc kubenswrapper[4796]: I0930 17:06:00.911600 4796 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 17:06:02 crc kubenswrapper[4796]: I0930 17:06:02.745273 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" path="/var/lib/kubelet/pods/90a2b4ab-61f5-49b9-9be8-aeab05d1ba58/volumes" Sep 30 17:06:04 crc kubenswrapper[4796]: I0930 17:06:04.825948 4796 scope.go:117] "RemoveContainer" containerID="e21a17e48af95794a5fb3b5abbd4975743723673ffb0e75c19b211ada2d08712" Sep 30 17:06:06 crc kubenswrapper[4796]: I0930 17:06:06.733543 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:06:07 crc kubenswrapper[4796]: I0930 17:06:07.777890 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerStarted","Data":"9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b"} Sep 30 17:06:10 crc kubenswrapper[4796]: I0930 17:06:10.734665 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:06:10 crc kubenswrapper[4796]: E0930 17:06:10.735082 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:06:11 crc kubenswrapper[4796]: I0930 17:06:11.625565 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:06:11 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:06:11 crc kubenswrapper[4796]: > Sep 30 17:06:13 crc kubenswrapper[4796]: I0930 17:06:13.733684 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:06:13 crc kubenswrapper[4796]: E0930 17:06:13.734172 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:06:14 crc kubenswrapper[4796]: I0930 17:06:14.687451 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:06:14 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:06:14 crc kubenswrapper[4796]: > Sep 30 17:06:17 crc kubenswrapper[4796]: I0930 17:06:17.623452 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" probeResult="failure" output=< Sep 30 17:06:17 crc kubenswrapper[4796]: Process cloudkitty-proc not found Sep 30 17:06:17 crc kubenswrapper[4796]: > Sep 30 17:06:17 crc kubenswrapper[4796]: I0930 17:06:17.624213 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cloudkitty-proc-0" Sep 30 17:06:17 crc kubenswrapper[4796]: I0930 17:06:17.625363 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cloudkitty-proc" containerStatusID={"Type":"cri-o","ID":"9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b"} pod="openstack/cloudkitty-proc-0" containerMessage="Container cloudkitty-proc failed liveness probe, will be restarted" Sep 30 17:06:17 crc kubenswrapper[4796]: I0930 17:06:17.625425 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerName="cloudkitty-proc" containerID="cri-o://9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" gracePeriod=30 Sep 30 17:06:18 crc kubenswrapper[4796]: E0930 17:06:18.611037 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:06:18 crc kubenswrapper[4796]: I0930 17:06:18.946445 4796 generic.go:334] "Generic (PLEG): container finished" podID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" exitCode=0 Sep 30 17:06:18 crc kubenswrapper[4796]: I0930 17:06:18.946506 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad","Type":"ContainerDied","Data":"9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b"} Sep 30 17:06:18 crc kubenswrapper[4796]: I0930 17:06:18.946556 4796 scope.go:117] "RemoveContainer" containerID="c9eb6e0d713f89b22432e1cf686173c2e915d936d79b367cbcbe53244d4bcd41" Sep 30 17:06:18 crc kubenswrapper[4796]: I0930 17:06:18.947595 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:06:18 crc kubenswrapper[4796]: E0930 17:06:18.947942 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:06:25 crc kubenswrapper[4796]: I0930 17:06:25.733582 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:06:25 crc kubenswrapper[4796]: E0930 17:06:25.734103 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:06:26 crc kubenswrapper[4796]: I0930 17:06:26.734023 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:06:26 crc kubenswrapper[4796]: E0930 17:06:26.734385 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:06:30 crc kubenswrapper[4796]: I0930 17:06:30.733899 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:06:30 crc kubenswrapper[4796]: E0930 17:06:30.734972 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:06:31 crc kubenswrapper[4796]: I0930 17:06:31.097014 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 17:06:31 crc kubenswrapper[4796]: I0930 17:06:31.097128 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.643548 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g66h7"] Sep 30 17:06:36 crc kubenswrapper[4796]: E0930 17:06:36.645123 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="extract-content" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.645206 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="extract-content" Sep 30 17:06:36 crc kubenswrapper[4796]: E0930 17:06:36.645300 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="gather" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.645355 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="gather" Sep 30 17:06:36 crc kubenswrapper[4796]: E0930 17:06:36.645411 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="registry-server" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.645462 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="registry-server" Sep 30 17:06:36 crc kubenswrapper[4796]: E0930 17:06:36.645517 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="copy" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.645569 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="copy" Sep 30 17:06:36 crc kubenswrapper[4796]: E0930 17:06:36.645633 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="extract-utilities" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.645688 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="extract-utilities" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.645959 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="copy" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.646094 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d104c70-9f39-4075-82a0-d6102d512746" containerName="registry-server" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.646175 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="90a2b4ab-61f5-49b9-9be8-aeab05d1ba58" containerName="gather" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.647889 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.677619 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g66h7"] Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.705297 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw5tt\" (UniqueName: \"kubernetes.io/projected/0020074b-bd6e-4b94-8794-b98cf6bc0f80-kube-api-access-dw5tt\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.705396 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-utilities\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.705461 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-catalog-content\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.807898 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-catalog-content\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.808132 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw5tt\" (UniqueName: \"kubernetes.io/projected/0020074b-bd6e-4b94-8794-b98cf6bc0f80-kube-api-access-dw5tt\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.808253 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-utilities\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.808441 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-catalog-content\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.809604 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-utilities\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.862383 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw5tt\" (UniqueName: \"kubernetes.io/projected/0020074b-bd6e-4b94-8794-b98cf6bc0f80-kube-api-access-dw5tt\") pod \"community-operators-g66h7\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:36 crc kubenswrapper[4796]: I0930 17:06:36.990663 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:37 crc kubenswrapper[4796]: I0930 17:06:37.348585 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g66h7"] Sep 30 17:06:37 crc kubenswrapper[4796]: W0930 17:06:37.356947 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0020074b_bd6e_4b94_8794_b98cf6bc0f80.slice/crio-daeda38327cc24d09cd5f830108e5b23fa605b56e914a2072e8cbc40fd5ad2b8 WatchSource:0}: Error finding container daeda38327cc24d09cd5f830108e5b23fa605b56e914a2072e8cbc40fd5ad2b8: Status 404 returned error can't find the container with id daeda38327cc24d09cd5f830108e5b23fa605b56e914a2072e8cbc40fd5ad2b8 Sep 30 17:06:38 crc kubenswrapper[4796]: I0930 17:06:38.215097 4796 generic.go:334] "Generic (PLEG): container finished" podID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerID="46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02" exitCode=0 Sep 30 17:06:38 crc kubenswrapper[4796]: I0930 17:06:38.215299 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerDied","Data":"46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02"} Sep 30 17:06:38 crc kubenswrapper[4796]: I0930 17:06:38.215494 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerStarted","Data":"daeda38327cc24d09cd5f830108e5b23fa605b56e914a2072e8cbc40fd5ad2b8"} Sep 30 17:06:38 crc kubenswrapper[4796]: I0930 17:06:38.753546 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:06:38 crc kubenswrapper[4796]: E0930 17:06:38.753795 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:06:39 crc kubenswrapper[4796]: I0930 17:06:39.733489 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:06:39 crc kubenswrapper[4796]: E0930 17:06:39.734206 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:06:40 crc kubenswrapper[4796]: I0930 17:06:40.243037 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerStarted","Data":"4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7"} Sep 30 17:06:43 crc kubenswrapper[4796]: I0930 17:06:43.283560 4796 generic.go:334] "Generic (PLEG): container finished" podID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerID="4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7" exitCode=0 Sep 30 17:06:43 crc kubenswrapper[4796]: I0930 17:06:43.283674 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerDied","Data":"4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7"} Sep 30 17:06:43 crc kubenswrapper[4796]: I0930 17:06:43.733254 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:06:43 crc kubenswrapper[4796]: E0930 17:06:43.733545 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:06:44 crc kubenswrapper[4796]: I0930 17:06:44.296238 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerStarted","Data":"d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14"} Sep 30 17:06:44 crc kubenswrapper[4796]: I0930 17:06:44.322951 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g66h7" podStartSLOduration=2.58969205 podStartE2EDuration="8.322925386s" podCreationTimestamp="2025-09-30 17:06:36 +0000 UTC" firstStartedPulling="2025-09-30 17:06:38.217825201 +0000 UTC m=+3290.231103728" lastFinishedPulling="2025-09-30 17:06:43.951058527 +0000 UTC m=+3295.964337064" observedRunningTime="2025-09-30 17:06:44.315729558 +0000 UTC m=+3296.329008095" watchObservedRunningTime="2025-09-30 17:06:44.322925386 +0000 UTC m=+3296.336203913" Sep 30 17:06:46 crc kubenswrapper[4796]: I0930 17:06:46.991874 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:46 crc kubenswrapper[4796]: I0930 17:06:46.992472 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:47 crc kubenswrapper[4796]: I0930 17:06:47.065357 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:52 crc kubenswrapper[4796]: I0930 17:06:52.734374 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:06:52 crc kubenswrapper[4796]: E0930 17:06:52.735456 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:06:54 crc kubenswrapper[4796]: I0930 17:06:54.738131 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:06:54 crc kubenswrapper[4796]: E0930 17:06:54.738784 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:06:57 crc kubenswrapper[4796]: I0930 17:06:57.198392 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:57 crc kubenswrapper[4796]: I0930 17:06:57.239681 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g66h7"] Sep 30 17:06:57 crc kubenswrapper[4796]: I0930 17:06:57.445736 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g66h7" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="registry-server" containerID="cri-o://d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14" gracePeriod=2 Sep 30 17:06:57 crc kubenswrapper[4796]: I0930 17:06:57.733956 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:06:57 crc kubenswrapper[4796]: E0930 17:06:57.734368 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:06:57 crc kubenswrapper[4796]: I0930 17:06:57.965279 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.034728 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw5tt\" (UniqueName: \"kubernetes.io/projected/0020074b-bd6e-4b94-8794-b98cf6bc0f80-kube-api-access-dw5tt\") pod \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.035015 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-utilities\") pod \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.035066 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-catalog-content\") pod \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\" (UID: \"0020074b-bd6e-4b94-8794-b98cf6bc0f80\") " Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.040776 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-utilities" (OuterVolumeSpecName: "utilities") pod "0020074b-bd6e-4b94-8794-b98cf6bc0f80" (UID: "0020074b-bd6e-4b94-8794-b98cf6bc0f80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.052313 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0020074b-bd6e-4b94-8794-b98cf6bc0f80-kube-api-access-dw5tt" (OuterVolumeSpecName: "kube-api-access-dw5tt") pod "0020074b-bd6e-4b94-8794-b98cf6bc0f80" (UID: "0020074b-bd6e-4b94-8794-b98cf6bc0f80"). InnerVolumeSpecName "kube-api-access-dw5tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.087246 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0020074b-bd6e-4b94-8794-b98cf6bc0f80" (UID: "0020074b-bd6e-4b94-8794-b98cf6bc0f80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.137381 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw5tt\" (UniqueName: \"kubernetes.io/projected/0020074b-bd6e-4b94-8794-b98cf6bc0f80-kube-api-access-dw5tt\") on node \"crc\" DevicePath \"\"" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.137421 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.137431 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020074b-bd6e-4b94-8794-b98cf6bc0f80-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.482620 4796 generic.go:334] "Generic (PLEG): container finished" podID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerID="d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14" exitCode=0 Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.482679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerDied","Data":"d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14"} Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.483038 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g66h7" event={"ID":"0020074b-bd6e-4b94-8794-b98cf6bc0f80","Type":"ContainerDied","Data":"daeda38327cc24d09cd5f830108e5b23fa605b56e914a2072e8cbc40fd5ad2b8"} Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.483070 4796 scope.go:117] "RemoveContainer" containerID="d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.482749 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g66h7" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.519665 4796 scope.go:117] "RemoveContainer" containerID="4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.528029 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g66h7"] Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.546863 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g66h7"] Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.556520 4796 scope.go:117] "RemoveContainer" containerID="46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.614268 4796 scope.go:117] "RemoveContainer" containerID="d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14" Sep 30 17:06:58 crc kubenswrapper[4796]: E0930 17:06:58.614827 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14\": container with ID starting with d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14 not found: ID does not exist" containerID="d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.614878 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14"} err="failed to get container status \"d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14\": rpc error: code = NotFound desc = could not find container \"d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14\": container with ID starting with d0999753c7752e3564d0f122367dbdc1898680a9bcaf33e2d7785d0cc0c61a14 not found: ID does not exist" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.614909 4796 scope.go:117] "RemoveContainer" containerID="4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7" Sep 30 17:06:58 crc kubenswrapper[4796]: E0930 17:06:58.615444 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7\": container with ID starting with 4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7 not found: ID does not exist" containerID="4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.615484 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7"} err="failed to get container status \"4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7\": rpc error: code = NotFound desc = could not find container \"4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7\": container with ID starting with 4da3fc3d9a866889643e3adfc1d7aeae709d05d8cc70349211b85958ba48b4c7 not found: ID does not exist" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.615511 4796 scope.go:117] "RemoveContainer" containerID="46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02" Sep 30 17:06:58 crc kubenswrapper[4796]: E0930 17:06:58.615934 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02\": container with ID starting with 46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02 not found: ID does not exist" containerID="46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.616032 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02"} err="failed to get container status \"46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02\": rpc error: code = NotFound desc = could not find container \"46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02\": container with ID starting with 46e815066c490a03af0ce77783ce9a55723e88a2fa8f88fe161d7ae2b4cd0f02 not found: ID does not exist" Sep 30 17:06:58 crc kubenswrapper[4796]: I0930 17:06:58.764562 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" path="/var/lib/kubelet/pods/0020074b-bd6e-4b94-8794-b98cf6bc0f80/volumes" Sep 30 17:07:01 crc kubenswrapper[4796]: I0930 17:07:01.096667 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 17:07:01 crc kubenswrapper[4796]: I0930 17:07:01.096752 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.608820 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wmwvc"] Sep 30 17:07:06 crc kubenswrapper[4796]: E0930 17:07:06.610589 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="registry-server" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.610610 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="registry-server" Sep 30 17:07:06 crc kubenswrapper[4796]: E0930 17:07:06.610627 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="extract-utilities" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.610634 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="extract-utilities" Sep 30 17:07:06 crc kubenswrapper[4796]: E0930 17:07:06.610678 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="extract-content" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.610684 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="extract-content" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.610877 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0020074b-bd6e-4b94-8794-b98cf6bc0f80" containerName="registry-server" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.612848 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.630290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-catalog-content\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.630363 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2m4\" (UniqueName: \"kubernetes.io/projected/1fe413f2-a70a-4d88-996d-52c63c964155-kube-api-access-bs2m4\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.632245 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-utilities\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.634637 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wmwvc"] Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.733476 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:07:06 crc kubenswrapper[4796]: E0930 17:07:06.733941 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.734716 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-utilities\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.734790 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-catalog-content\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.734845 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2m4\" (UniqueName: \"kubernetes.io/projected/1fe413f2-a70a-4d88-996d-52c63c964155-kube-api-access-bs2m4\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.735715 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-utilities\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.736044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-catalog-content\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.758084 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2m4\" (UniqueName: \"kubernetes.io/projected/1fe413f2-a70a-4d88-996d-52c63c964155-kube-api-access-bs2m4\") pod \"redhat-operators-wmwvc\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:06 crc kubenswrapper[4796]: I0930 17:07:06.941245 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:07 crc kubenswrapper[4796]: I0930 17:07:07.384685 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wmwvc"] Sep 30 17:07:07 crc kubenswrapper[4796]: I0930 17:07:07.612362 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerStarted","Data":"d3c8496a19732dbf003c5fcb97fd48618ff63ec68601bdb6c24e98ae0e998a2a"} Sep 30 17:07:08 crc kubenswrapper[4796]: I0930 17:07:08.626746 4796 generic.go:334] "Generic (PLEG): container finished" podID="1fe413f2-a70a-4d88-996d-52c63c964155" containerID="a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b" exitCode=0 Sep 30 17:07:08 crc kubenswrapper[4796]: I0930 17:07:08.626870 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerDied","Data":"a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b"} Sep 30 17:07:08 crc kubenswrapper[4796]: I0930 17:07:08.741733 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:07:08 crc kubenswrapper[4796]: E0930 17:07:08.742300 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:07:09 crc kubenswrapper[4796]: I0930 17:07:09.734589 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:07:09 crc kubenswrapper[4796]: E0930 17:07:09.735787 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:07:10 crc kubenswrapper[4796]: I0930 17:07:10.653387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerStarted","Data":"29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8"} Sep 30 17:07:12 crc kubenswrapper[4796]: I0930 17:07:12.678155 4796 generic.go:334] "Generic (PLEG): container finished" podID="1fe413f2-a70a-4d88-996d-52c63c964155" containerID="29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8" exitCode=0 Sep 30 17:07:12 crc kubenswrapper[4796]: I0930 17:07:12.678445 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerDied","Data":"29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8"} Sep 30 17:07:13 crc kubenswrapper[4796]: I0930 17:07:13.691063 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerStarted","Data":"61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b"} Sep 30 17:07:13 crc kubenswrapper[4796]: I0930 17:07:13.712954 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wmwvc" podStartSLOduration=3.212210994 podStartE2EDuration="7.712932646s" podCreationTimestamp="2025-09-30 17:07:06 +0000 UTC" firstStartedPulling="2025-09-30 17:07:08.629822066 +0000 UTC m=+3320.643100593" lastFinishedPulling="2025-09-30 17:07:13.130543678 +0000 UTC m=+3325.143822245" observedRunningTime="2025-09-30 17:07:13.707254422 +0000 UTC m=+3325.720532969" watchObservedRunningTime="2025-09-30 17:07:13.712932646 +0000 UTC m=+3325.726211173" Sep 30 17:07:16 crc kubenswrapper[4796]: I0930 17:07:16.942226 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:16 crc kubenswrapper[4796]: I0930 17:07:16.942897 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:18 crc kubenswrapper[4796]: I0930 17:07:18.007327 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wmwvc" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="registry-server" probeResult="failure" output=< Sep 30 17:07:18 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Sep 30 17:07:18 crc kubenswrapper[4796]: > Sep 30 17:07:21 crc kubenswrapper[4796]: I0930 17:07:21.734128 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:07:21 crc kubenswrapper[4796]: E0930 17:07:21.734809 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:07:22 crc kubenswrapper[4796]: I0930 17:07:22.733196 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:07:22 crc kubenswrapper[4796]: E0930 17:07:22.733513 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:07:23 crc kubenswrapper[4796]: I0930 17:07:23.733361 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:07:23 crc kubenswrapper[4796]: E0930 17:07:23.734184 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:07:27 crc kubenswrapper[4796]: I0930 17:07:27.016766 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:27 crc kubenswrapper[4796]: I0930 17:07:27.075554 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:27 crc kubenswrapper[4796]: I0930 17:07:27.262291 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wmwvc"] Sep 30 17:07:28 crc kubenswrapper[4796]: I0930 17:07:28.881871 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wmwvc" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="registry-server" containerID="cri-o://61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b" gracePeriod=2 Sep 30 17:07:29 crc kubenswrapper[4796]: E0930 17:07:29.090323 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fe413f2_a70a_4d88_996d_52c63c964155.slice/crio-conmon-61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b.scope\": RecentStats: unable to find data in memory cache]" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.435641 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.548556 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs2m4\" (UniqueName: \"kubernetes.io/projected/1fe413f2-a70a-4d88-996d-52c63c964155-kube-api-access-bs2m4\") pod \"1fe413f2-a70a-4d88-996d-52c63c964155\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.548657 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-catalog-content\") pod \"1fe413f2-a70a-4d88-996d-52c63c964155\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.548743 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-utilities\") pod \"1fe413f2-a70a-4d88-996d-52c63c964155\" (UID: \"1fe413f2-a70a-4d88-996d-52c63c964155\") " Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.550338 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-utilities" (OuterVolumeSpecName: "utilities") pod "1fe413f2-a70a-4d88-996d-52c63c964155" (UID: "1fe413f2-a70a-4d88-996d-52c63c964155"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.555241 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fe413f2-a70a-4d88-996d-52c63c964155-kube-api-access-bs2m4" (OuterVolumeSpecName: "kube-api-access-bs2m4") pod "1fe413f2-a70a-4d88-996d-52c63c964155" (UID: "1fe413f2-a70a-4d88-996d-52c63c964155"). InnerVolumeSpecName "kube-api-access-bs2m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.630542 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fe413f2-a70a-4d88-996d-52c63c964155" (UID: "1fe413f2-a70a-4d88-996d-52c63c964155"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.651792 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs2m4\" (UniqueName: \"kubernetes.io/projected/1fe413f2-a70a-4d88-996d-52c63c964155-kube-api-access-bs2m4\") on node \"crc\" DevicePath \"\"" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.652511 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.652708 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fe413f2-a70a-4d88-996d-52c63c964155-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.897289 4796 generic.go:334] "Generic (PLEG): container finished" podID="1fe413f2-a70a-4d88-996d-52c63c964155" containerID="61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b" exitCode=0 Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.897328 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerDied","Data":"61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b"} Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.897356 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmwvc" event={"ID":"1fe413f2-a70a-4d88-996d-52c63c964155","Type":"ContainerDied","Data":"d3c8496a19732dbf003c5fcb97fd48618ff63ec68601bdb6c24e98ae0e998a2a"} Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.897372 4796 scope.go:117] "RemoveContainer" containerID="61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.899055 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmwvc" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.956659 4796 scope.go:117] "RemoveContainer" containerID="29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8" Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.979536 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wmwvc"] Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.986514 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wmwvc"] Sep 30 17:07:29 crc kubenswrapper[4796]: I0930 17:07:29.995531 4796 scope.go:117] "RemoveContainer" containerID="a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.046175 4796 scope.go:117] "RemoveContainer" containerID="61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b" Sep 30 17:07:30 crc kubenswrapper[4796]: E0930 17:07:30.046648 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b\": container with ID starting with 61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b not found: ID does not exist" containerID="61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.046704 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b"} err="failed to get container status \"61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b\": rpc error: code = NotFound desc = could not find container \"61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b\": container with ID starting with 61de72cb2324229740ff74b9fb05698574c904e000615ade07bb857fa56de69b not found: ID does not exist" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.046739 4796 scope.go:117] "RemoveContainer" containerID="29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8" Sep 30 17:07:30 crc kubenswrapper[4796]: E0930 17:07:30.047722 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8\": container with ID starting with 29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8 not found: ID does not exist" containerID="29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.047773 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8"} err="failed to get container status \"29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8\": rpc error: code = NotFound desc = could not find container \"29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8\": container with ID starting with 29def3e58ab71b4f7d1b8f493a51710aa3d0b429eb83834dce4eacb2e9b2a8e8 not found: ID does not exist" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.047800 4796 scope.go:117] "RemoveContainer" containerID="a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b" Sep 30 17:07:30 crc kubenswrapper[4796]: E0930 17:07:30.048072 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b\": container with ID starting with a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b not found: ID does not exist" containerID="a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.048101 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b"} err="failed to get container status \"a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b\": rpc error: code = NotFound desc = could not find container \"a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b\": container with ID starting with a445df54fe4efeae8e130bf517ce0138aefa51dc41fd6c2b71e4e1a606ae7f3b not found: ID does not exist" Sep 30 17:07:30 crc kubenswrapper[4796]: I0930 17:07:30.758532 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" path="/var/lib/kubelet/pods/1fe413f2-a70a-4d88-996d-52c63c964155/volumes" Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.097261 4796 patch_prober.go:28] interesting pod/machine-config-daemon-8bbws container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.097349 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.097425 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.098487 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0e16983f6a64c78a8ee2d7ba623c69e9a76d02d91fe4467bdd766d749b98a1d0"} pod="openshift-machine-config-operator/machine-config-daemon-8bbws" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.098612 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" podUID="670c655e-3953-4773-84ef-19c678d482f9" containerName="machine-config-daemon" containerID="cri-o://0e16983f6a64c78a8ee2d7ba623c69e9a76d02d91fe4467bdd766d749b98a1d0" gracePeriod=600 Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.921562 4796 generic.go:334] "Generic (PLEG): container finished" podID="670c655e-3953-4773-84ef-19c678d482f9" containerID="0e16983f6a64c78a8ee2d7ba623c69e9a76d02d91fe4467bdd766d749b98a1d0" exitCode=0 Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.921630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerDied","Data":"0e16983f6a64c78a8ee2d7ba623c69e9a76d02d91fe4467bdd766d749b98a1d0"} Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.921887 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8bbws" event={"ID":"670c655e-3953-4773-84ef-19c678d482f9","Type":"ContainerStarted","Data":"a545835a3ce8e058ff27ff48b6c1ddccbba6f5a70df5556384a2224c8d01dc5f"} Sep 30 17:07:31 crc kubenswrapper[4796]: I0930 17:07:31.921909 4796 scope.go:117] "RemoveContainer" containerID="263c242250659e11b0abb4542855ff33ed6236d1ea3353196e46ebd0f7da5708" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.678068 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xqcjt"] Sep 30 17:07:33 crc kubenswrapper[4796]: E0930 17:07:33.678834 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="extract-utilities" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.678850 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="extract-utilities" Sep 30 17:07:33 crc kubenswrapper[4796]: E0930 17:07:33.678870 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="registry-server" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.678878 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="registry-server" Sep 30 17:07:33 crc kubenswrapper[4796]: E0930 17:07:33.678934 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="extract-content" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.678942 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="extract-content" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.679232 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fe413f2-a70a-4d88-996d-52c63c964155" containerName="registry-server" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.681328 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.693073 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqcjt"] Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.734764 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:07:33 crc kubenswrapper[4796]: E0930 17:07:33.735015 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.755094 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-utilities\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.755255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7sqz\" (UniqueName: \"kubernetes.io/projected/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-kube-api-access-r7sqz\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.755510 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-catalog-content\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.857279 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7sqz\" (UniqueName: \"kubernetes.io/projected/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-kube-api-access-r7sqz\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.857523 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-catalog-content\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.857678 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-utilities\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.859483 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-catalog-content\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.859618 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-utilities\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:33 crc kubenswrapper[4796]: I0930 17:07:33.883650 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7sqz\" (UniqueName: \"kubernetes.io/projected/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-kube-api-access-r7sqz\") pod \"redhat-marketplace-xqcjt\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:34 crc kubenswrapper[4796]: I0930 17:07:34.007669 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:34 crc kubenswrapper[4796]: I0930 17:07:34.480694 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqcjt"] Sep 30 17:07:34 crc kubenswrapper[4796]: W0930 17:07:34.481178 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddba5b4df_f2bf_4e81_8bc0_c458d543a4c9.slice/crio-2b9e1900c9b3991ff54dfdd94643504829d06d7cf3265f679d9f9f26c0fe786f WatchSource:0}: Error finding container 2b9e1900c9b3991ff54dfdd94643504829d06d7cf3265f679d9f9f26c0fe786f: Status 404 returned error can't find the container with id 2b9e1900c9b3991ff54dfdd94643504829d06d7cf3265f679d9f9f26c0fe786f Sep 30 17:07:34 crc kubenswrapper[4796]: I0930 17:07:34.977397 4796 generic.go:334] "Generic (PLEG): container finished" podID="dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" containerID="9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6" exitCode=0 Sep 30 17:07:34 crc kubenswrapper[4796]: I0930 17:07:34.977801 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerDied","Data":"9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6"} Sep 30 17:07:34 crc kubenswrapper[4796]: I0930 17:07:34.977840 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerStarted","Data":"2b9e1900c9b3991ff54dfdd94643504829d06d7cf3265f679d9f9f26c0fe786f"} Sep 30 17:07:35 crc kubenswrapper[4796]: I0930 17:07:35.999595 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerStarted","Data":"ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535"} Sep 30 17:07:36 crc kubenswrapper[4796]: I0930 17:07:36.733826 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:07:36 crc kubenswrapper[4796]: E0930 17:07:36.734696 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:07:37 crc kubenswrapper[4796]: I0930 17:07:37.011313 4796 generic.go:334] "Generic (PLEG): container finished" podID="dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" containerID="ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535" exitCode=0 Sep 30 17:07:37 crc kubenswrapper[4796]: I0930 17:07:37.011627 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerDied","Data":"ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535"} Sep 30 17:07:38 crc kubenswrapper[4796]: I0930 17:07:38.024438 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerStarted","Data":"1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6"} Sep 30 17:07:38 crc kubenswrapper[4796]: I0930 17:07:38.053241 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xqcjt" podStartSLOduration=2.367015817 podStartE2EDuration="5.05321927s" podCreationTimestamp="2025-09-30 17:07:33 +0000 UTC" firstStartedPulling="2025-09-30 17:07:34.980076533 +0000 UTC m=+3346.993355060" lastFinishedPulling="2025-09-30 17:07:37.666279946 +0000 UTC m=+3349.679558513" observedRunningTime="2025-09-30 17:07:38.049470361 +0000 UTC m=+3350.062748888" watchObservedRunningTime="2025-09-30 17:07:38.05321927 +0000 UTC m=+3350.066497797" Sep 30 17:07:38 crc kubenswrapper[4796]: I0930 17:07:38.744652 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:07:38 crc kubenswrapper[4796]: E0930 17:07:38.744904 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:07:44 crc kubenswrapper[4796]: I0930 17:07:44.008583 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:44 crc kubenswrapper[4796]: I0930 17:07:44.009289 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:44 crc kubenswrapper[4796]: I0930 17:07:44.064069 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:44 crc kubenswrapper[4796]: I0930 17:07:44.167839 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:45 crc kubenswrapper[4796]: I0930 17:07:45.657743 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqcjt"] Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.120213 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xqcjt" podUID="dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" containerName="registry-server" containerID="cri-o://1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6" gracePeriod=2 Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.699625 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.737998 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7sqz\" (UniqueName: \"kubernetes.io/projected/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-kube-api-access-r7sqz\") pod \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.738135 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-utilities\") pod \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.738181 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-catalog-content\") pod \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\" (UID: \"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9\") " Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.740851 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-utilities" (OuterVolumeSpecName: "utilities") pod "dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" (UID: "dba5b4df-f2bf-4e81-8bc0-c458d543a4c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.751464 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-kube-api-access-r7sqz" (OuterVolumeSpecName: "kube-api-access-r7sqz") pod "dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" (UID: "dba5b4df-f2bf-4e81-8bc0-c458d543a4c9"). InnerVolumeSpecName "kube-api-access-r7sqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.765883 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" (UID: "dba5b4df-f2bf-4e81-8bc0-c458d543a4c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.842312 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7sqz\" (UniqueName: \"kubernetes.io/projected/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-kube-api-access-r7sqz\") on node \"crc\" DevicePath \"\"" Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.843102 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 17:07:46 crc kubenswrapper[4796]: I0930 17:07:46.843247 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.138296 4796 generic.go:334] "Generic (PLEG): container finished" podID="dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" containerID="1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6" exitCode=0 Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.138356 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerDied","Data":"1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6"} Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.138382 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqcjt" event={"ID":"dba5b4df-f2bf-4e81-8bc0-c458d543a4c9","Type":"ContainerDied","Data":"2b9e1900c9b3991ff54dfdd94643504829d06d7cf3265f679d9f9f26c0fe786f"} Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.138400 4796 scope.go:117] "RemoveContainer" containerID="1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.138405 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqcjt" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.164821 4796 scope.go:117] "RemoveContainer" containerID="ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.190416 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqcjt"] Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.200141 4796 scope.go:117] "RemoveContainer" containerID="9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.229899 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqcjt"] Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.245371 4796 scope.go:117] "RemoveContainer" containerID="1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6" Sep 30 17:07:47 crc kubenswrapper[4796]: E0930 17:07:47.246280 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6\": container with ID starting with 1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6 not found: ID does not exist" containerID="1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.246314 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6"} err="failed to get container status \"1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6\": rpc error: code = NotFound desc = could not find container \"1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6\": container with ID starting with 1aedc9d5e2f03d235cf368ede6e60721dcb0007003967a9177e3639ba43027c6 not found: ID does not exist" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.246335 4796 scope.go:117] "RemoveContainer" containerID="ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535" Sep 30 17:07:47 crc kubenswrapper[4796]: E0930 17:07:47.246891 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535\": container with ID starting with ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535 not found: ID does not exist" containerID="ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.246941 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535"} err="failed to get container status \"ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535\": rpc error: code = NotFound desc = could not find container \"ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535\": container with ID starting with ffcfa39ddd85bf4f7cb16221e3a713110366173bfdc1f5b64f254038a287d535 not found: ID does not exist" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.246974 4796 scope.go:117] "RemoveContainer" containerID="9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6" Sep 30 17:07:47 crc kubenswrapper[4796]: E0930 17:07:47.247468 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6\": container with ID starting with 9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6 not found: ID does not exist" containerID="9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.247494 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6"} err="failed to get container status \"9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6\": rpc error: code = NotFound desc = could not find container \"9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6\": container with ID starting with 9c422052109c538b07f5b63a8e52e8715ff6b5d268ec5cd4afd051e63f176ab6 not found: ID does not exist" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.733423 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:07:47 crc kubenswrapper[4796]: I0930 17:07:47.734780 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:07:47 crc kubenswrapper[4796]: E0930 17:07:47.735368 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:07:47 crc kubenswrapper[4796]: E0930 17:07:47.735913 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:07:48 crc kubenswrapper[4796]: I0930 17:07:48.752792 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dba5b4df-f2bf-4e81-8bc0-c458d543a4c9" path="/var/lib/kubelet/pods/dba5b4df-f2bf-4e81-8bc0-c458d543a4c9/volumes" Sep 30 17:07:53 crc kubenswrapper[4796]: I0930 17:07:53.733668 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:07:53 crc kubenswrapper[4796]: E0930 17:07:53.734799 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:07:59 crc kubenswrapper[4796]: I0930 17:07:59.733472 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:07:59 crc kubenswrapper[4796]: E0930 17:07:59.734802 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:08:01 crc kubenswrapper[4796]: I0930 17:08:01.733173 4796 scope.go:117] "RemoveContainer" containerID="60289bd391515b90d56c6210cbc4236931b802200149c62490dd9aa3ce1d3e7a" Sep 30 17:08:01 crc kubenswrapper[4796]: E0930 17:08:01.733688 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-cpv5x_openstack(165eb10b-4c60-4635-928e-c04a72e9b4f5)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-cpv5x" podUID="165eb10b-4c60-4635-928e-c04a72e9b4f5" Sep 30 17:08:05 crc kubenswrapper[4796]: I0930 17:08:05.003522 4796 scope.go:117] "RemoveContainer" containerID="f8834f992b2ad1927b79ab593c9628bf18c6eddf805e59cfb79faa0c38eb55b5" Sep 30 17:08:08 crc kubenswrapper[4796]: I0930 17:08:08.746631 4796 scope.go:117] "RemoveContainer" containerID="9996520da919f2247a442dbc8db0c68027d11e9f9eeb0163c9f243d08c21315b" Sep 30 17:08:08 crc kubenswrapper[4796]: E0930 17:08:08.747678 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-proc\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cloudkitty-proc pod=cloudkitty-proc-0_openstack(1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad)\"" pod="openstack/cloudkitty-proc-0" podUID="1dd26425-9aaf-4a1a-be8e-1fbfd2d248ad" Sep 30 17:08:10 crc kubenswrapper[4796]: I0930 17:08:10.733683 4796 scope.go:117] "RemoveContainer" containerID="e3c4aea781e4711a1320b702a2c27e62489a10d5b61f70e9cd0b10d4b3116e59" Sep 30 17:08:10 crc kubenswrapper[4796]: E0930 17:08:10.734795 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gateway pod=cloudkitty-lokistack-gateway-89dc74b89-hv6jn_openstack(15e69566-5b25-4929-a40b-1a3e2d710800)\"" pod="openstack/cloudkitty-lokistack-gateway-89dc74b89-hv6jn" podUID="15e69566-5b25-4929-a40b-1a3e2d710800" Sep 30 17:08:11 crc kubenswrapper[4796]: I0930 17:08:11.060835 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-rpkc5"] Sep 30 17:08:11 crc kubenswrapper[4796]: I0930 17:08:11.076972 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-rpkc5"] var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067007201024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067007202017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067000160016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067000160015450 5ustar corecore